var/home/core/zuul-output/0000755000175000017500000000000015113404455014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015113417000015461 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005102167115113416772017707 0ustar rootrootDec 01 21:34:40 crc systemd[1]: Starting Kubernetes Kubelet... Dec 01 21:34:40 crc restorecon[4695]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 21:34:41 crc restorecon[4695]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 21:34:41 crc restorecon[4695]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 01 21:34:41 crc kubenswrapper[4857]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 21:34:41 crc kubenswrapper[4857]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 01 21:34:41 crc kubenswrapper[4857]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 21:34:41 crc kubenswrapper[4857]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 21:34:41 crc kubenswrapper[4857]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 01 21:34:41 crc kubenswrapper[4857]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.659085 4857 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664019 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664067 4857 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664074 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664082 4857 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664089 4857 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664097 4857 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664106 4857 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664114 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664121 4857 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664127 4857 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664134 4857 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664139 4857 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664144 4857 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664151 4857 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664158 4857 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664165 4857 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664170 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664176 4857 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664186 4857 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664192 4857 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664199 4857 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664204 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664210 4857 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664216 4857 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664221 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664227 4857 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664233 4857 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664238 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664243 4857 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664249 4857 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664254 4857 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664260 4857 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664265 4857 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664271 4857 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664279 4857 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664286 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664294 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664299 4857 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664305 4857 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664311 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664317 4857 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664322 4857 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664328 4857 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664333 4857 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664338 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664343 4857 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664348 4857 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664355 4857 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664362 4857 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664368 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664374 4857 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664380 4857 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664386 4857 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664391 4857 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664397 4857 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664403 4857 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664408 4857 feature_gate.go:330] unrecognized feature gate: Example Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664415 4857 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664421 4857 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664427 4857 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664433 4857 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664439 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664445 4857 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664452 4857 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664457 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664462 4857 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664468 4857 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664473 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664478 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664483 4857 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.664489 4857 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664762 4857 flags.go:64] FLAG: --address="0.0.0.0" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664777 4857 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664789 4857 flags.go:64] FLAG: --anonymous-auth="true" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664797 4857 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664805 4857 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664811 4857 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664819 4857 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664827 4857 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664834 4857 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664840 4857 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664846 4857 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664852 4857 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664859 4857 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664864 4857 flags.go:64] FLAG: --cgroup-root="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664870 4857 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664876 4857 flags.go:64] FLAG: --client-ca-file="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664882 4857 flags.go:64] FLAG: --cloud-config="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664888 4857 flags.go:64] FLAG: --cloud-provider="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664894 4857 flags.go:64] FLAG: --cluster-dns="[]" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664901 4857 flags.go:64] FLAG: --cluster-domain="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664907 4857 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664914 4857 flags.go:64] FLAG: --config-dir="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664920 4857 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664926 4857 flags.go:64] FLAG: --container-log-max-files="5" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664934 4857 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664941 4857 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664947 4857 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664953 4857 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664959 4857 flags.go:64] FLAG: --contention-profiling="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664965 4857 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664971 4857 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664978 4857 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664984 4857 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664992 4857 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.664998 4857 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665004 4857 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665010 4857 flags.go:64] FLAG: --enable-load-reader="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665017 4857 flags.go:64] FLAG: --enable-server="true" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665023 4857 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665033 4857 flags.go:64] FLAG: --event-burst="100" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665058 4857 flags.go:64] FLAG: --event-qps="50" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665065 4857 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665072 4857 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665080 4857 flags.go:64] FLAG: --eviction-hard="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665089 4857 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665096 4857 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665102 4857 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665109 4857 flags.go:64] FLAG: --eviction-soft="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665115 4857 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665121 4857 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665127 4857 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665134 4857 flags.go:64] FLAG: --experimental-mounter-path="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665140 4857 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665146 4857 flags.go:64] FLAG: --fail-swap-on="true" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665152 4857 flags.go:64] FLAG: --feature-gates="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665164 4857 flags.go:64] FLAG: --file-check-frequency="20s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665170 4857 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665176 4857 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665183 4857 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665189 4857 flags.go:64] FLAG: --healthz-port="10248" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665196 4857 flags.go:64] FLAG: --help="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665202 4857 flags.go:64] FLAG: --hostname-override="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665208 4857 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665214 4857 flags.go:64] FLAG: --http-check-frequency="20s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665222 4857 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665229 4857 flags.go:64] FLAG: --image-credential-provider-config="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665235 4857 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665241 4857 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665247 4857 flags.go:64] FLAG: --image-service-endpoint="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665253 4857 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665259 4857 flags.go:64] FLAG: --kube-api-burst="100" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665265 4857 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665271 4857 flags.go:64] FLAG: --kube-api-qps="50" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665278 4857 flags.go:64] FLAG: --kube-reserved="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665284 4857 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665290 4857 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665296 4857 flags.go:64] FLAG: --kubelet-cgroups="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665302 4857 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665308 4857 flags.go:64] FLAG: --lock-file="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665313 4857 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665320 4857 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665326 4857 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665334 4857 flags.go:64] FLAG: --log-json-split-stream="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665340 4857 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665348 4857 flags.go:64] FLAG: --log-text-split-stream="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665354 4857 flags.go:64] FLAG: --logging-format="text" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665360 4857 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665366 4857 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665372 4857 flags.go:64] FLAG: --manifest-url="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665378 4857 flags.go:64] FLAG: --manifest-url-header="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665386 4857 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665392 4857 flags.go:64] FLAG: --max-open-files="1000000" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665400 4857 flags.go:64] FLAG: --max-pods="110" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665406 4857 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665412 4857 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665418 4857 flags.go:64] FLAG: --memory-manager-policy="None" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665425 4857 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665431 4857 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665437 4857 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665443 4857 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665458 4857 flags.go:64] FLAG: --node-status-max-images="50" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665464 4857 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665470 4857 flags.go:64] FLAG: --oom-score-adj="-999" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665476 4857 flags.go:64] FLAG: --pod-cidr="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665481 4857 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665491 4857 flags.go:64] FLAG: --pod-manifest-path="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665497 4857 flags.go:64] FLAG: --pod-max-pids="-1" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665503 4857 flags.go:64] FLAG: --pods-per-core="0" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665509 4857 flags.go:64] FLAG: --port="10250" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665516 4857 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665522 4857 flags.go:64] FLAG: --provider-id="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665528 4857 flags.go:64] FLAG: --qos-reserved="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665534 4857 flags.go:64] FLAG: --read-only-port="10255" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665540 4857 flags.go:64] FLAG: --register-node="true" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665546 4857 flags.go:64] FLAG: --register-schedulable="true" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665552 4857 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665562 4857 flags.go:64] FLAG: --registry-burst="10" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665568 4857 flags.go:64] FLAG: --registry-qps="5" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665574 4857 flags.go:64] FLAG: --reserved-cpus="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665580 4857 flags.go:64] FLAG: --reserved-memory="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665587 4857 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665593 4857 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665599 4857 flags.go:64] FLAG: --rotate-certificates="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665606 4857 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665612 4857 flags.go:64] FLAG: --runonce="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665618 4857 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665624 4857 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665630 4857 flags.go:64] FLAG: --seccomp-default="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665636 4857 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665642 4857 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665649 4857 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665655 4857 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665662 4857 flags.go:64] FLAG: --storage-driver-password="root" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665668 4857 flags.go:64] FLAG: --storage-driver-secure="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665673 4857 flags.go:64] FLAG: --storage-driver-table="stats" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665679 4857 flags.go:64] FLAG: --storage-driver-user="root" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665685 4857 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665692 4857 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665698 4857 flags.go:64] FLAG: --system-cgroups="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665704 4857 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665713 4857 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665719 4857 flags.go:64] FLAG: --tls-cert-file="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665726 4857 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665734 4857 flags.go:64] FLAG: --tls-min-version="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665741 4857 flags.go:64] FLAG: --tls-private-key-file="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665748 4857 flags.go:64] FLAG: --topology-manager-policy="none" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665754 4857 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665760 4857 flags.go:64] FLAG: --topology-manager-scope="container" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665767 4857 flags.go:64] FLAG: --v="2" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665774 4857 flags.go:64] FLAG: --version="false" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665782 4857 flags.go:64] FLAG: --vmodule="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665788 4857 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.665795 4857 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.665978 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.665990 4857 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.665998 4857 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666005 4857 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666011 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666019 4857 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666026 4857 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666033 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666065 4857 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666073 4857 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666081 4857 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666087 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666093 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666099 4857 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666104 4857 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666109 4857 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666117 4857 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666124 4857 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666131 4857 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666138 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666144 4857 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666150 4857 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666156 4857 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666161 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666167 4857 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666172 4857 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666177 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666183 4857 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666193 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666200 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666207 4857 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666214 4857 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666221 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666227 4857 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666234 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666240 4857 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666247 4857 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666254 4857 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666261 4857 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666269 4857 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666278 4857 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666285 4857 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666291 4857 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666300 4857 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666308 4857 feature_gate.go:330] unrecognized feature gate: Example Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666314 4857 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666321 4857 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666328 4857 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666335 4857 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666342 4857 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666349 4857 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666356 4857 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666364 4857 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666370 4857 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666380 4857 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666387 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666394 4857 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666400 4857 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666408 4857 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666416 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666424 4857 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666430 4857 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666436 4857 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666441 4857 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666448 4857 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666453 4857 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666459 4857 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666464 4857 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666470 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666475 4857 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.666480 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.666494 4857 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.677336 4857 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.677381 4857 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677492 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677504 4857 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677512 4857 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677519 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677527 4857 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677534 4857 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677540 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677545 4857 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677552 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677558 4857 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677563 4857 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677569 4857 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677575 4857 feature_gate.go:330] unrecognized feature gate: Example Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677580 4857 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677586 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677593 4857 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677598 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677604 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677610 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677617 4857 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677622 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677628 4857 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677638 4857 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677645 4857 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677652 4857 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677659 4857 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677665 4857 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677672 4857 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677679 4857 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677686 4857 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677693 4857 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677700 4857 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677706 4857 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677713 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677721 4857 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677728 4857 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677733 4857 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677740 4857 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677745 4857 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677750 4857 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677756 4857 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677761 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677767 4857 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677772 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677777 4857 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677783 4857 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677789 4857 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677794 4857 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677799 4857 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677808 4857 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677816 4857 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677822 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677828 4857 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677836 4857 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677842 4857 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677849 4857 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677854 4857 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677862 4857 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677867 4857 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677873 4857 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677879 4857 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677885 4857 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677890 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677896 4857 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677901 4857 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677906 4857 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677914 4857 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677920 4857 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677926 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677933 4857 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.677939 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.677949 4857 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678140 4857 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678151 4857 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678157 4857 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678163 4857 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678170 4857 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678176 4857 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678182 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678187 4857 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678192 4857 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678198 4857 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678203 4857 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678211 4857 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678217 4857 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678224 4857 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678230 4857 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678236 4857 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678242 4857 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678247 4857 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678254 4857 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678260 4857 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678266 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678271 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678278 4857 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678285 4857 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678291 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678297 4857 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678303 4857 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678308 4857 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678314 4857 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678319 4857 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678325 4857 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678331 4857 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678336 4857 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678342 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678347 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678353 4857 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678359 4857 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678365 4857 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678371 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678376 4857 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678382 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678388 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678394 4857 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678401 4857 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678408 4857 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678416 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678422 4857 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678428 4857 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678434 4857 feature_gate.go:330] unrecognized feature gate: Example Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678440 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678447 4857 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678453 4857 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678459 4857 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678464 4857 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678470 4857 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678476 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678482 4857 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678487 4857 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678492 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678498 4857 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678503 4857 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678508 4857 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678514 4857 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678519 4857 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678524 4857 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678530 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678535 4857 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678541 4857 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678546 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678552 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.678557 4857 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.678566 4857 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.678973 4857 server.go:940] "Client rotation is on, will bootstrap in background" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.682716 4857 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.682815 4857 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.683495 4857 server.go:997] "Starting client certificate rotation" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.683542 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.683713 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-28 11:56:30.184477531 +0000 UTC Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.683810 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.689668 4857 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.691760 4857 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 21:34:41 crc kubenswrapper[4857]: E1201 21:34:41.692192 4857 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.702644 4857 log.go:25] "Validated CRI v1 runtime API" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.727877 4857 log.go:25] "Validated CRI v1 image API" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.732916 4857 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.736218 4857 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-01-21-30-31-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.736320 4857 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.760814 4857 manager.go:217] Machine: {Timestamp:2025-12-01 21:34:41.75911686 +0000 UTC m=+0.249179217 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:0d498ed1-bf21-4e30-b0f7-bd8e4f32b853 BootID:ea7fffe2-6d6d-498a-b750-d72f9c51be69 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:91:dd:f3 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:91:dd:f3 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ae:35:bf Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:dd:10:c1 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:68:31:e2 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:49:9a:8f Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ee:a3:9b:03:8b:59 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:6a:ee:61:2b:10:b8 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.761129 4857 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.761308 4857 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.762108 4857 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.762333 4857 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.762374 4857 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.762623 4857 topology_manager.go:138] "Creating topology manager with none policy" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.762636 4857 container_manager_linux.go:303] "Creating device plugin manager" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.762893 4857 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.762936 4857 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.763384 4857 state_mem.go:36] "Initialized new in-memory state store" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.763508 4857 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.764705 4857 kubelet.go:418] "Attempting to sync node with API server" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.764727 4857 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.764755 4857 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.764773 4857 kubelet.go:324] "Adding apiserver pod source" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.764796 4857 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.767501 4857 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.767730 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Dec 01 21:34:41 crc kubenswrapper[4857]: E1201 21:34:41.767826 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.767915 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Dec 01 21:34:41 crc kubenswrapper[4857]: E1201 21:34:41.768094 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.768726 4857 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.772229 4857 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773116 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773168 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773187 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773205 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773233 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773251 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773269 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773298 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773320 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773339 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773391 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773410 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.773708 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.774599 4857 server.go:1280] "Started kubelet" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.774714 4857 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.774914 4857 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.775677 4857 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.776038 4857 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 01 21:34:41 crc systemd[1]: Started Kubernetes Kubelet. Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.777034 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.777110 4857 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.777421 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 08:58:10.230686164 +0000 UTC Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.777533 4857 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.777570 4857 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 01 21:34:41 crc kubenswrapper[4857]: E1201 21:34:41.777273 4857 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.136:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d34fb07fca2c9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 21:34:41.774535369 +0000 UTC m=+0.264597726,LastTimestamp:2025-12-01 21:34:41.774535369 +0000 UTC m=+0.264597726,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.777786 4857 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 01 21:34:41 crc kubenswrapper[4857]: E1201 21:34:41.777567 4857 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.778434 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.778526 4857 factory.go:55] Registering systemd factory Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.778549 4857 factory.go:221] Registration of the systemd container factory successfully Dec 01 21:34:41 crc kubenswrapper[4857]: E1201 21:34:41.778563 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.778958 4857 factory.go:153] Registering CRI-O factory Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.778979 4857 factory.go:221] Registration of the crio container factory successfully Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.779194 4857 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.779226 4857 factory.go:103] Registering Raw factory Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.779251 4857 manager.go:1196] Started watching for new ooms in manager Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.780221 4857 manager.go:319] Starting recovery of all containers Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.781121 4857 server.go:460] "Adding debug handlers to kubelet server" Dec 01 21:34:41 crc kubenswrapper[4857]: E1201 21:34:41.783564 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="200ms" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793185 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793686 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793721 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793747 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793762 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793779 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793800 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793817 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793841 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793862 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793906 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793927 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793943 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793975 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.793996 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794014 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794029 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794071 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794092 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794133 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794147 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794162 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794183 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794198 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794214 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794236 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794260 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794277 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794299 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794316 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794337 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794352 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794366 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794383 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794403 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794417 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794437 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794462 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794484 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794498 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794511 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794532 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794546 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794564 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794577 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794593 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794610 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794628 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794647 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794666 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794680 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794697 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794718 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794798 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794824 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.794842 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799578 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799630 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799652 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799673 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799694 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799714 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799734 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799755 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799774 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799793 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799839 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799857 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799878 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799913 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799932 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799954 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.799975 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800022 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800070 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800097 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800124 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800142 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800163 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800184 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800203 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800222 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800241 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800261 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800279 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800299 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800318 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800336 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800355 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800374 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800392 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800410 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800428 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800448 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800469 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800491 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800512 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800532 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800551 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800576 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800597 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800615 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800634 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800653 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800720 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800753 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800781 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800831 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800853 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800873 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800896 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800916 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800937 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800958 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800978 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.800997 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.801019 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802619 4857 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802699 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802730 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802757 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802778 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802797 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802815 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802835 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802857 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802878 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802898 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802917 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802938 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802957 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802976 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.802996 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803015 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803034 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803100 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803122 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803143 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803164 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803183 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803201 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803218 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803238 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803267 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803286 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803307 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803328 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803348 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803366 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803424 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803445 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803467 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803489 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803508 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803528 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803547 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803567 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803587 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803609 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803630 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803651 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803673 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803694 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803716 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803735 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803754 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803773 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803792 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803815 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803835 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803855 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803876 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803897 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803918 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803937 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803957 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.803977 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804029 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804070 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804093 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804114 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804134 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804154 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804172 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804195 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804216 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804239 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804259 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804277 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804298 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804317 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804335 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804355 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804374 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804395 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804414 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804435 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804453 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804472 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804492 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804515 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804533 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804553 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804570 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804591 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804611 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804629 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804650 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804670 4857 reconstruct.go:97] "Volume reconstruction finished" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.804682 4857 reconciler.go:26] "Reconciler: start to sync state" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.811810 4857 manager.go:324] Recovery completed Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.823672 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.827751 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.827822 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.827834 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.829247 4857 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.829282 4857 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.829308 4857 state_mem.go:36] "Initialized new in-memory state store" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.831383 4857 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.833333 4857 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.833484 4857 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.833585 4857 kubelet.go:2335] "Starting kubelet main sync loop" Dec 01 21:34:41 crc kubenswrapper[4857]: E1201 21:34:41.833713 4857 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 01 21:34:41 crc kubenswrapper[4857]: W1201 21:34:41.834672 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Dec 01 21:34:41 crc kubenswrapper[4857]: E1201 21:34:41.834790 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.839664 4857 policy_none.go:49] "None policy: Start" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.841240 4857 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.841273 4857 state_mem.go:35] "Initializing new in-memory state store" Dec 01 21:34:41 crc kubenswrapper[4857]: E1201 21:34:41.878284 4857 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.900836 4857 manager.go:334] "Starting Device Plugin manager" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.901180 4857 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.901204 4857 server.go:79] "Starting device plugin registration server" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.901745 4857 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.901769 4857 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.901923 4857 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.902072 4857 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.902087 4857 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 01 21:34:41 crc kubenswrapper[4857]: E1201 21:34:41.912763 4857 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.934443 4857 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.934581 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.936215 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.936271 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.936284 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.936536 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.936802 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.936855 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.937758 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.937787 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.937799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.937900 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.938107 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.938145 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.938333 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.938365 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.938375 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.938674 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.938701 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.938715 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.938825 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.939250 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.939560 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.939574 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.939589 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.939601 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.940188 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.940222 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.940236 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.940354 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.940461 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.940495 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.941668 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.941688 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.941697 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.941827 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.941845 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.941855 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.941874 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.941903 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.941935 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.941992 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.942005 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.942534 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.942554 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:41 crc kubenswrapper[4857]: I1201 21:34:41.942569 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:41 crc kubenswrapper[4857]: E1201 21:34:41.984517 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="400ms" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.004402 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.005774 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.005817 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.005826 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.005852 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 21:34:42 crc kubenswrapper[4857]: E1201 21:34:42.006333 4857 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007448 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007496 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007536 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007570 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007601 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007629 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007659 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007686 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007714 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007744 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007797 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007865 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007895 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007924 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.007953 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109399 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109476 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109513 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109539 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109567 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109595 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109628 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109660 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109693 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109706 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109757 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109688 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109770 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109843 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109737 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109726 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109804 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109820 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109773 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.109926 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.110109 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.110178 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.110222 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.110270 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.110274 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.110305 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.110349 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.110411 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.110454 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.110495 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.207334 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.209423 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.209489 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.209513 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.209553 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 21:34:42 crc kubenswrapper[4857]: E1201 21:34:42.210508 4857 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.259335 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.275505 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: W1201 21:34:42.296985 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-14dafe4dade09d0d1ad649337b8df8ca4483effa85c84a7cbe6ea2bedc307fbf WatchSource:0}: Error finding container 14dafe4dade09d0d1ad649337b8df8ca4483effa85c84a7cbe6ea2bedc307fbf: Status 404 returned error can't find the container with id 14dafe4dade09d0d1ad649337b8df8ca4483effa85c84a7cbe6ea2bedc307fbf Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.297740 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: W1201 21:34:42.307527 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e62c1a8be2f92652fdee209a36f362ff40c6e7201728da8468c235155844b62c WatchSource:0}: Error finding container e62c1a8be2f92652fdee209a36f362ff40c6e7201728da8468c235155844b62c: Status 404 returned error can't find the container with id e62c1a8be2f92652fdee209a36f362ff40c6e7201728da8468c235155844b62c Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.313430 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.319341 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 21:34:42 crc kubenswrapper[4857]: W1201 21:34:42.322785 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-fcbd09b86aaa28c12b3756f35c80c23a1fdc1983279a0dd74e57e8bca6a8fc87 WatchSource:0}: Error finding container fcbd09b86aaa28c12b3756f35c80c23a1fdc1983279a0dd74e57e8bca6a8fc87: Status 404 returned error can't find the container with id fcbd09b86aaa28c12b3756f35c80c23a1fdc1983279a0dd74e57e8bca6a8fc87 Dec 01 21:34:42 crc kubenswrapper[4857]: W1201 21:34:42.327528 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b928af0742f1f2f29912abc358bbfbe2261340f6aeb19c75bf593c19c2e64165 WatchSource:0}: Error finding container b928af0742f1f2f29912abc358bbfbe2261340f6aeb19c75bf593c19c2e64165: Status 404 returned error can't find the container with id b928af0742f1f2f29912abc358bbfbe2261340f6aeb19c75bf593c19c2e64165 Dec 01 21:34:42 crc kubenswrapper[4857]: W1201 21:34:42.370355 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-09e032578781765bdd160ef780904028438dc9f4d1cac4ec0e3ddd70ab53077a WatchSource:0}: Error finding container 09e032578781765bdd160ef780904028438dc9f4d1cac4ec0e3ddd70ab53077a: Status 404 returned error can't find the container with id 09e032578781765bdd160ef780904028438dc9f4d1cac4ec0e3ddd70ab53077a Dec 01 21:34:42 crc kubenswrapper[4857]: E1201 21:34:42.386133 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="800ms" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.611021 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.612158 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.612192 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.612202 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.612222 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 21:34:42 crc kubenswrapper[4857]: E1201 21:34:42.612607 4857 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Dec 01 21:34:42 crc kubenswrapper[4857]: W1201 21:34:42.688925 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Dec 01 21:34:42 crc kubenswrapper[4857]: E1201 21:34:42.688998 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Dec 01 21:34:42 crc kubenswrapper[4857]: W1201 21:34:42.748769 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Dec 01 21:34:42 crc kubenswrapper[4857]: E1201 21:34:42.748872 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.776473 4857 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.777643 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 11:57:28.051660267 +0000 UTC Dec 01 21:34:42 crc kubenswrapper[4857]: W1201 21:34:42.836295 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Dec 01 21:34:42 crc kubenswrapper[4857]: E1201 21:34:42.836387 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.840346 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc"} Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.840445 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e62c1a8be2f92652fdee209a36f362ff40c6e7201728da8468c235155844b62c"} Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.842561 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230" exitCode=0 Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.842580 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230"} Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.842605 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"14dafe4dade09d0d1ad649337b8df8ca4483effa85c84a7cbe6ea2bedc307fbf"} Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.842707 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.843698 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.843737 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.843751 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.844293 4857 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="14266fb1f3b756d929206f0879d19d488e5a00a58397aa24faa068fde565faf0" exitCode=0 Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.844360 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"14266fb1f3b756d929206f0879d19d488e5a00a58397aa24faa068fde565faf0"} Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.844387 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"09e032578781765bdd160ef780904028438dc9f4d1cac4ec0e3ddd70ab53077a"} Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.844486 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.845251 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.845409 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.845443 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.845455 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.846085 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.846115 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.846126 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.846348 4857 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394" exitCode=0 Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.846412 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394"} Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.846441 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b928af0742f1f2f29912abc358bbfbe2261340f6aeb19c75bf593c19c2e64165"} Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.846508 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.849641 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.849663 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.849672 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.850530 4857 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192" exitCode=0 Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.850581 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192"} Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.850733 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.850949 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"fcbd09b86aaa28c12b3756f35c80c23a1fdc1983279a0dd74e57e8bca6a8fc87"} Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.851741 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.851766 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:42 crc kubenswrapper[4857]: I1201 21:34:42.851775 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:43 crc kubenswrapper[4857]: W1201 21:34:43.186448 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Dec 01 21:34:43 crc kubenswrapper[4857]: E1201 21:34:43.186532 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Dec 01 21:34:43 crc kubenswrapper[4857]: E1201 21:34:43.186892 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="1.6s" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.412913 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.416004 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.416303 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.416363 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.416404 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 21:34:43 crc kubenswrapper[4857]: E1201 21:34:43.416874 4857 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.778326 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 10:59:28.473949428 +0000 UTC Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.778662 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 517h24m44.695289946s for next certificate rotation Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.835667 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.857157 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188"} Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.857210 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03"} Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.857225 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1"} Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.857324 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.858562 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.858598 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.858611 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.860816 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.860803 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"005b5e00ddaba85767652a371cfdae9b98238fc5b1ec89beecdb2c18d9b2062f"} Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.866007 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.866036 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.866057 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.868546 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7"} Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.868596 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101"} Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.868611 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98"} Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.868680 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.871315 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.871374 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.871390 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.873293 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14"} Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.873322 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091"} Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.873337 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda"} Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.873349 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16"} Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.875184 4857 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7b5fee847e910929ba7c528edbfa8d3cb9d9a9b401ffee73c06007a25e9d83fa" exitCode=0 Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.875224 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7b5fee847e910929ba7c528edbfa8d3cb9d9a9b401ffee73c06007a25e9d83fa"} Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.875329 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.875977 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.876010 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:43 crc kubenswrapper[4857]: I1201 21:34:43.876023 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.882950 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773"} Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.882981 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.884759 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.884811 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.884835 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.887221 4857 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="db87a1360dfeebc1019dafd80edfe0c1606e73b03423c95c3dfbd41bed205a30" exitCode=0 Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.887301 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.887316 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"db87a1360dfeebc1019dafd80edfe0c1606e73b03423c95c3dfbd41bed205a30"} Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.887350 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.887514 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.888007 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.888029 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.888030 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.888059 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.888064 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.888075 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.889730 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.889772 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:44 crc kubenswrapper[4857]: I1201 21:34:44.889788 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.017669 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.018836 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.018907 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.018926 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.018953 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.753807 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.775579 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.781017 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.895858 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"03510bb7637d1cbe32a9f556176f57b300da1ff8c1c7250a31dada11e62a2628"} Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.895933 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.895990 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.895932 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"aa9a8627cfc2ddd816242fca3f829623bd35d8809a7de1f7fb17754d58267d8e"} Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.896127 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.896137 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e1217723138f40bf6cc0f26b526966801027ad75fa2bd1a6b2cc17819f3626c5"} Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.896175 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"51eeb99bcca72b8b5dca0db8c91deb42958cd3891198894cb13f7f8add6b25ed"} Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.897025 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.897124 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.897145 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.897466 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.897539 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:45 crc kubenswrapper[4857]: I1201 21:34:45.897568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.905663 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3db0d155ba2f6824f6d9c6e5339f2f5ee503b62a8516d90e270914e7a12e1d7b"} Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.905721 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.905787 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.905812 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.905835 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.905853 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.907499 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.907545 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.907563 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.907648 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.907686 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.907706 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.907649 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.907775 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:46 crc kubenswrapper[4857]: I1201 21:34:46.907799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:47 crc kubenswrapper[4857]: I1201 21:34:47.831848 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 01 21:34:47 crc kubenswrapper[4857]: I1201 21:34:47.862489 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:47 crc kubenswrapper[4857]: I1201 21:34:47.908299 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:47 crc kubenswrapper[4857]: I1201 21:34:47.908407 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:47 crc kubenswrapper[4857]: I1201 21:34:47.909551 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:47 crc kubenswrapper[4857]: I1201 21:34:47.909624 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:47 crc kubenswrapper[4857]: I1201 21:34:47.909644 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:47 crc kubenswrapper[4857]: I1201 21:34:47.910236 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:47 crc kubenswrapper[4857]: I1201 21:34:47.910289 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:47 crc kubenswrapper[4857]: I1201 21:34:47.910313 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:48 crc kubenswrapper[4857]: I1201 21:34:48.257369 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:48 crc kubenswrapper[4857]: I1201 21:34:48.257599 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:34:48 crc kubenswrapper[4857]: I1201 21:34:48.257664 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:48 crc kubenswrapper[4857]: I1201 21:34:48.259295 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:48 crc kubenswrapper[4857]: I1201 21:34:48.259384 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:48 crc kubenswrapper[4857]: I1201 21:34:48.259409 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:48 crc kubenswrapper[4857]: I1201 21:34:48.833691 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 01 21:34:48 crc kubenswrapper[4857]: I1201 21:34:48.910955 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:48 crc kubenswrapper[4857]: I1201 21:34:48.912308 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:48 crc kubenswrapper[4857]: I1201 21:34:48.912361 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:48 crc kubenswrapper[4857]: I1201 21:34:48.912380 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.396472 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.396681 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.398281 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.398329 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.398348 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.831229 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.831480 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.832938 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.832993 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.833009 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.913083 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.914231 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.914281 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:49 crc kubenswrapper[4857]: I1201 21:34:49.914298 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:51 crc kubenswrapper[4857]: E1201 21:34:51.913585 4857 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 21:34:52 crc kubenswrapper[4857]: I1201 21:34:52.109210 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:52 crc kubenswrapper[4857]: I1201 21:34:52.109472 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:52 crc kubenswrapper[4857]: I1201 21:34:52.110795 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:52 crc kubenswrapper[4857]: I1201 21:34:52.110830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:52 crc kubenswrapper[4857]: I1201 21:34:52.110841 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:52 crc kubenswrapper[4857]: I1201 21:34:52.115701 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:52 crc kubenswrapper[4857]: I1201 21:34:52.200717 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:34:52 crc kubenswrapper[4857]: I1201 21:34:52.921278 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:52 crc kubenswrapper[4857]: I1201 21:34:52.922192 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:52 crc kubenswrapper[4857]: I1201 21:34:52.922223 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:52 crc kubenswrapper[4857]: I1201 21:34:52.922235 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:53 crc kubenswrapper[4857]: I1201 21:34:53.778347 4857 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 01 21:34:53 crc kubenswrapper[4857]: E1201 21:34:53.837673 4857 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 21:34:53 crc kubenswrapper[4857]: I1201 21:34:53.924649 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:53 crc kubenswrapper[4857]: I1201 21:34:53.925976 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:53 crc kubenswrapper[4857]: I1201 21:34:53.926023 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:53 crc kubenswrapper[4857]: I1201 21:34:53.926034 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:54 crc kubenswrapper[4857]: W1201 21:34:54.387794 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 01 21:34:54 crc kubenswrapper[4857]: I1201 21:34:54.387984 4857 trace.go:236] Trace[1557620946]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 21:34:44.386) (total time: 10001ms): Dec 01 21:34:54 crc kubenswrapper[4857]: Trace[1557620946]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (21:34:54.387) Dec 01 21:34:54 crc kubenswrapper[4857]: Trace[1557620946]: [10.00191174s] [10.00191174s] END Dec 01 21:34:54 crc kubenswrapper[4857]: E1201 21:34:54.388028 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 21:34:54 crc kubenswrapper[4857]: W1201 21:34:54.550364 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 01 21:34:54 crc kubenswrapper[4857]: I1201 21:34:54.550554 4857 trace.go:236] Trace[1808565328]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 21:34:44.548) (total time: 10002ms): Dec 01 21:34:54 crc kubenswrapper[4857]: Trace[1808565328]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (21:34:54.550) Dec 01 21:34:54 crc kubenswrapper[4857]: Trace[1808565328]: [10.002248012s] [10.002248012s] END Dec 01 21:34:54 crc kubenswrapper[4857]: E1201 21:34:54.550604 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 21:34:54 crc kubenswrapper[4857]: E1201 21:34:54.788213 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 01 21:34:55 crc kubenswrapper[4857]: E1201 21:34:55.020474 4857 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 01 21:34:55 crc kubenswrapper[4857]: W1201 21:34:55.044827 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 01 21:34:55 crc kubenswrapper[4857]: I1201 21:34:55.044957 4857 trace.go:236] Trace[1774887498]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 21:34:45.043) (total time: 10001ms): Dec 01 21:34:55 crc kubenswrapper[4857]: Trace[1774887498]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (21:34:55.044) Dec 01 21:34:55 crc kubenswrapper[4857]: Trace[1774887498]: [10.001358005s] [10.001358005s] END Dec 01 21:34:55 crc kubenswrapper[4857]: E1201 21:34:55.044992 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 21:34:55 crc kubenswrapper[4857]: I1201 21:34:55.082708 4857 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 21:34:55 crc kubenswrapper[4857]: I1201 21:34:55.082796 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 21:34:55 crc kubenswrapper[4857]: I1201 21:34:55.097127 4857 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 21:34:55 crc kubenswrapper[4857]: I1201 21:34:55.097200 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 21:34:55 crc kubenswrapper[4857]: I1201 21:34:55.200884 4857 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 21:34:55 crc kubenswrapper[4857]: I1201 21:34:55.201228 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 21:34:57 crc kubenswrapper[4857]: I1201 21:34:57.872021 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 01 21:34:57 crc kubenswrapper[4857]: I1201 21:34:57.872802 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:57 crc kubenswrapper[4857]: I1201 21:34:57.874454 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:57 crc kubenswrapper[4857]: I1201 21:34:57.874514 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:57 crc kubenswrapper[4857]: I1201 21:34:57.874539 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:57 crc kubenswrapper[4857]: I1201 21:34:57.890849 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 01 21:34:57 crc kubenswrapper[4857]: I1201 21:34:57.935983 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:57 crc kubenswrapper[4857]: I1201 21:34:57.937003 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:57 crc kubenswrapper[4857]: I1201 21:34:57.937233 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:57 crc kubenswrapper[4857]: I1201 21:34:57.937384 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.195499 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.214997 4857 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.220606 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.222508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.222597 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.222617 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.222669 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 21:34:58 crc kubenswrapper[4857]: E1201 21:34:58.227979 4857 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.265501 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.266286 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.267974 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.268016 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.268030 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.271765 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.340544 4857 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.479640 4857 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.773676 4857 apiserver.go:52] "Watching apiserver" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.779144 4857 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.779676 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.780215 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.780851 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.780891 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.780929 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.780933 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.780851 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 21:34:58 crc kubenswrapper[4857]: E1201 21:34:58.782069 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:34:58 crc kubenswrapper[4857]: E1201 21:34:58.782322 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:34:58 crc kubenswrapper[4857]: E1201 21:34:58.782429 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.783112 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.783263 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.783420 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.783804 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.784638 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.784638 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.784753 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.784816 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.785412 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.821072 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.838796 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.862546 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.879085 4857 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.881033 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.896900 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.910607 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.927966 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:34:58 crc kubenswrapper[4857]: I1201 21:34:58.956483 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 21:34:59 crc kubenswrapper[4857]: I1201 21:34:59.833987 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:34:59 crc kubenswrapper[4857]: E1201 21:34:59.834187 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:34:59 crc kubenswrapper[4857]: I1201 21:34:59.940617 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.088427 4857 trace.go:236] Trace[710801613]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 21:34:45.968) (total time: 14120ms): Dec 01 21:35:00 crc kubenswrapper[4857]: Trace[710801613]: ---"Objects listed" error: 14120ms (21:35:00.088) Dec 01 21:35:00 crc kubenswrapper[4857]: Trace[710801613]: [14.120356309s] [14.120356309s] END Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.088469 4857 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.088760 4857 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.166598 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.182455 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189570 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189623 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189648 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189676 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189702 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189725 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189748 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189775 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189800 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189826 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189850 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189874 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.189989 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190014 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190018 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190070 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190034 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190133 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190156 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190173 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190189 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190209 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190224 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190287 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190305 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190320 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190083 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190026 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190955 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190074 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190284 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191002 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191212 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190787 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.190922 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191241 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191282 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191312 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191343 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191461 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191467 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191550 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191565 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191584 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191610 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191635 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191660 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191690 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191765 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191929 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191655 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191723 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191806 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191796 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.191978 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.192107 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:35:00.691882395 +0000 UTC m=+19.181944732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.195173 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.195214 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.195267 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.195431 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.195554 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.192193 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.192271 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.195650 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.192307 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.192330 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.195750 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.195779 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.195810 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.192580 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.192847 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.192905 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.193546 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.193713 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.194799 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.195665 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.196065 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.196104 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.196431 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.196731 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.197853 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.198380 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.212660 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.212747 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.212922 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.212966 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.213097 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.213090 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.213196 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.214309 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.214317 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.214388 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.215646 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.215811 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.215864 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.215897 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.215928 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.215954 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.216170 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.216232 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.215979 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.216303 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.216353 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.216430 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.216546 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.216561 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.216586 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.216619 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.216630 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.218961 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.219000 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.219026 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.219069 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.219097 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.219123 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.219164 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.219187 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.219211 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.219235 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.219263 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.219288 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220371 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220410 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220431 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220487 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220507 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220524 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220539 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220555 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220572 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220590 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220607 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220623 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220640 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220674 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220691 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220709 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220726 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220744 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220760 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220777 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220793 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220809 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220875 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220895 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220914 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220933 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220975 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220993 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221011 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221027 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221067 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221086 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221103 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221123 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221142 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221159 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221175 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221191 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221211 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221239 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221257 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221274 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221290 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221306 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221323 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221426 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221444 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221465 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221483 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221506 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221983 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222018 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222063 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222088 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222112 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222136 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222165 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222192 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222225 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222257 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222284 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222310 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222335 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222363 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222388 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222727 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222757 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222783 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222808 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222829 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222847 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222864 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222882 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222899 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222918 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222936 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222952 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222968 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222987 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223004 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223022 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223055 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223072 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223088 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223104 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223121 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223138 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223162 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223177 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223194 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223212 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223228 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223248 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223265 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223281 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223298 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223314 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223329 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223345 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223363 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223379 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223395 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223414 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223432 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223452 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223469 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223489 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228441 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228540 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228591 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228617 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228635 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228656 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228686 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228707 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228731 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228758 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228776 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228791 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228813 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228830 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228850 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228866 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228885 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228902 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228919 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228963 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228991 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229010 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229033 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229067 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229089 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229110 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229128 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229146 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229165 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229183 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229200 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229218 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229237 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229300 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229313 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229323 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229362 4857 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229373 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229384 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229398 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229410 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229420 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229430 4857 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229439 4857 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229449 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229458 4857 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229467 4857 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229478 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229487 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229496 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229506 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229515 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229524 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229535 4857 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229544 4857 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229552 4857 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229565 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229574 4857 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229583 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229593 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229603 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229612 4857 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229622 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229632 4857 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229641 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229651 4857 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229661 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229671 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229681 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229690 4857 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229699 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229708 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229719 4857 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229729 4857 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229738 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229747 4857 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229756 4857 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229765 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229773 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229783 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229792 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229801 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.219591 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220099 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220283 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.220466 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221245 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.230083 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221363 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221650 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.221960 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222209 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.222380 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223359 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.223464 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.225624 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.226282 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.226535 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.226912 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.227021 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.227256 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.227334 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.227413 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.227509 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.227725 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228112 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.228649 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229412 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.229657 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.230264 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.230495 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.230945 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.231078 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.231152 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.231293 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.231503 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.231783 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.231960 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.232121 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.232141 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.232195 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.232402 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.232423 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.232425 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.232687 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.232895 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.232907 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.232957 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.233027 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.233162 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.233309 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.233447 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.233376 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.233688 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.233728 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.233757 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.233927 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.233941 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.234006 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.234348 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.234535 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.235024 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.235195 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.235234 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.235298 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.235555 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.235628 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.235734 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.235879 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.235997 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.236204 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.236400 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.236609 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.236786 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.236806 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.236999 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.237076 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.237225 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.237393 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.237497 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.237648 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.237741 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.237944 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.237981 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.238193 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.238373 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.238396 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.238550 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.238625 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.238718 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.238867 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.238867 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.238887 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.239108 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.239130 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.239323 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.239339 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.239380 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.239624 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.239749 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.240006 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.240243 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.240655 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.241295 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.241492 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.241649 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.241689 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.242175 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.242613 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.242993 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.243019 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.243452 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.243753 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.244028 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.244614 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.244947 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.245227 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.245520 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.245551 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.246132 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.246816 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.247022 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.247361 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.248183 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.248457 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.248558 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.248702 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.248755 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.248807 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.248844 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.248957 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.249055 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.249107 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:00.749090263 +0000 UTC m=+19.239152580 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.249385 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.249451 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:00.749431362 +0000 UTC m=+19.239493879 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.249845 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.249936 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.250614 4857 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.251237 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.251468 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.251722 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.252074 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.253112 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.253206 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.253225 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.253736 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.254613 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.255207 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.256405 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.259314 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.261706 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.265155 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.265189 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.265203 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.265258 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:00.765241191 +0000 UTC m=+19.255303508 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.267128 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.273458 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.273502 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.273518 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.273581 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:00.773559756 +0000 UTC m=+19.263622073 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.273684 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.273863 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.274201 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.276288 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.282585 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.283904 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.288783 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.290878 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.295213 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.304278 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.314704 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.321448 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332097 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332171 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332256 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332267 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332276 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332284 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332293 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332302 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332310 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332319 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332329 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332339 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332348 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332357 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332382 4857 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332391 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332399 4857 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332408 4857 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332416 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332424 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332432 4857 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332431 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332441 4857 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332476 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332485 4857 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332495 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332503 4857 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332511 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332519 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332527 4857 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332535 4857 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332543 4857 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332552 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332562 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332571 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332578 4857 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332607 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332617 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332624 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332634 4857 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332641 4857 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332649 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332656 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332664 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332672 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332680 4857 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332688 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332698 4857 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332706 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332715 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332723 4857 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332731 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332738 4857 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332746 4857 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332754 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332764 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332773 4857 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332782 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332790 4857 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332798 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332812 4857 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332819 4857 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332828 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332837 4857 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332844 4857 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332852 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332859 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332866 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332876 4857 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332884 4857 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332891 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332900 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332907 4857 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332920 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332929 4857 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332936 4857 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332944 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333003 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333012 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333020 4857 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333027 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333035 4857 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333053 4857 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333062 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333073 4857 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333081 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333089 4857 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333097 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333105 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333113 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333121 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333137 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333147 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333156 4857 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333163 4857 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333172 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333180 4857 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333189 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333198 4857 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333206 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333214 4857 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333222 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333230 4857 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333239 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333246 4857 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333255 4857 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333263 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333270 4857 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333278 4857 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333286 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333294 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333302 4857 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333310 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333318 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333326 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333334 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333342 4857 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333366 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333376 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333384 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333392 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333400 4857 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333409 4857 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333417 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333427 4857 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333435 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333442 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333450 4857 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333458 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333466 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333474 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333483 4857 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333491 4857 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333501 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333510 4857 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333519 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333528 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333536 4857 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333544 4857 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333551 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333559 4857 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333568 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333577 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333585 4857 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333593 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333601 4857 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.333608 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.332607 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.337369 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.338202 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.444947 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.481247 4857 csr.go:261] certificate signing request csr-2sv9w is approved, waiting to be issued Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.526417 4857 csr.go:257] certificate signing request csr-2sv9w is issued Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.603108 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 21:35:00 crc kubenswrapper[4857]: W1201 21:35:00.634869 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-00d1f56b24847a23cb70c6f7884c65a78d93eacba35b177205fdc506a51380c4 WatchSource:0}: Error finding container 00d1f56b24847a23cb70c6f7884c65a78d93eacba35b177205fdc506a51380c4: Status 404 returned error can't find the container with id 00d1f56b24847a23cb70c6f7884c65a78d93eacba35b177205fdc506a51380c4 Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.747754 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.747900 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:35:01.747876062 +0000 UTC m=+20.237938379 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.834234 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.834251 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.834364 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.834465 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.848383 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.848469 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.848500 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.848518 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.848600 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.848645 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.848694 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.848712 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.848725 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.848733 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:01.848699627 +0000 UTC m=+20.338762094 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.848761 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:01.848752718 +0000 UTC m=+20.338815255 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.848787 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:01.848769979 +0000 UTC m=+20.338832296 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.848840 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.848848 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.848856 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:00 crc kubenswrapper[4857]: E1201 21:35:00.848892 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:01.848884102 +0000 UTC m=+20.338946419 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.943981 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08"} Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.944019 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"00d1f56b24847a23cb70c6f7884c65a78d93eacba35b177205fdc506a51380c4"} Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.945977 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0"} Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.946279 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed"} Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.946290 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0c8ea19e5b03436e513f5c2b81befa6bf230fecdc2e3af3fe1959b899129d72c"} Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.948078 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2d1c2d029e44b139fbf086cf1653d601f79e2861a0f7880a7b2a67d3faa4d70f"} Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.954594 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.967105 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.977562 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:00 crc kubenswrapper[4857]: I1201 21:35:00.990595 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.001199 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.015288 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.027413 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.038229 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.057852 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.077913 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.088805 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.108410 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.108475 4857 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.133710 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.147630 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.312608 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-r7x8v"] Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.312953 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-r7x8v" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.313931 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-89689"] Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.314209 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.328513 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.329012 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.329111 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.352698 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.355873 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.363482 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f63b0e62-e9e8-432e-9a1b-f5be926fc2ce-hosts-file\") pod \"node-resolver-r7x8v\" (UID: \"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\") " pod="openshift-dns/node-resolver-r7x8v" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.363533 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ac5dfac0-8933-4db8-9cb6-5dca0537d066-rootfs\") pod \"machine-config-daemon-89689\" (UID: \"ac5dfac0-8933-4db8-9cb6-5dca0537d066\") " pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.363594 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ac5dfac0-8933-4db8-9cb6-5dca0537d066-proxy-tls\") pod \"machine-config-daemon-89689\" (UID: \"ac5dfac0-8933-4db8-9cb6-5dca0537d066\") " pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.363616 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bqvm\" (UniqueName: \"kubernetes.io/projected/ac5dfac0-8933-4db8-9cb6-5dca0537d066-kube-api-access-6bqvm\") pod \"machine-config-daemon-89689\" (UID: \"ac5dfac0-8933-4db8-9cb6-5dca0537d066\") " pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.363641 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2pv5\" (UniqueName: \"kubernetes.io/projected/f63b0e62-e9e8-432e-9a1b-f5be926fc2ce-kube-api-access-k2pv5\") pod \"node-resolver-r7x8v\" (UID: \"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\") " pod="openshift-dns/node-resolver-r7x8v" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.363662 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ac5dfac0-8933-4db8-9cb6-5dca0537d066-mcd-auth-proxy-config\") pod \"machine-config-daemon-89689\" (UID: \"ac5dfac0-8933-4db8-9cb6-5dca0537d066\") " pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.364928 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.365004 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.365203 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.450599 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.464958 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ac5dfac0-8933-4db8-9cb6-5dca0537d066-proxy-tls\") pod \"machine-config-daemon-89689\" (UID: \"ac5dfac0-8933-4db8-9cb6-5dca0537d066\") " pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.465000 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bqvm\" (UniqueName: \"kubernetes.io/projected/ac5dfac0-8933-4db8-9cb6-5dca0537d066-kube-api-access-6bqvm\") pod \"machine-config-daemon-89689\" (UID: \"ac5dfac0-8933-4db8-9cb6-5dca0537d066\") " pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.465023 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2pv5\" (UniqueName: \"kubernetes.io/projected/f63b0e62-e9e8-432e-9a1b-f5be926fc2ce-kube-api-access-k2pv5\") pod \"node-resolver-r7x8v\" (UID: \"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\") " pod="openshift-dns/node-resolver-r7x8v" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.465246 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ac5dfac0-8933-4db8-9cb6-5dca0537d066-mcd-auth-proxy-config\") pod \"machine-config-daemon-89689\" (UID: \"ac5dfac0-8933-4db8-9cb6-5dca0537d066\") " pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.465300 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f63b0e62-e9e8-432e-9a1b-f5be926fc2ce-hosts-file\") pod \"node-resolver-r7x8v\" (UID: \"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\") " pod="openshift-dns/node-resolver-r7x8v" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.465320 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ac5dfac0-8933-4db8-9cb6-5dca0537d066-rootfs\") pod \"machine-config-daemon-89689\" (UID: \"ac5dfac0-8933-4db8-9cb6-5dca0537d066\") " pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.465379 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ac5dfac0-8933-4db8-9cb6-5dca0537d066-rootfs\") pod \"machine-config-daemon-89689\" (UID: \"ac5dfac0-8933-4db8-9cb6-5dca0537d066\") " pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.467459 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f63b0e62-e9e8-432e-9a1b-f5be926fc2ce-hosts-file\") pod \"node-resolver-r7x8v\" (UID: \"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\") " pod="openshift-dns/node-resolver-r7x8v" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.467951 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ac5dfac0-8933-4db8-9cb6-5dca0537d066-mcd-auth-proxy-config\") pod \"machine-config-daemon-89689\" (UID: \"ac5dfac0-8933-4db8-9cb6-5dca0537d066\") " pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.471585 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ac5dfac0-8933-4db8-9cb6-5dca0537d066-proxy-tls\") pod \"machine-config-daemon-89689\" (UID: \"ac5dfac0-8933-4db8-9cb6-5dca0537d066\") " pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.508015 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bqvm\" (UniqueName: \"kubernetes.io/projected/ac5dfac0-8933-4db8-9cb6-5dca0537d066-kube-api-access-6bqvm\") pod \"machine-config-daemon-89689\" (UID: \"ac5dfac0-8933-4db8-9cb6-5dca0537d066\") " pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.527184 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-01 21:30:00 +0000 UTC, rotation deadline is 2026-09-01 06:26:08.173972181 +0000 UTC Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.527273 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6560h51m6.64670148s for next certificate rotation Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.546418 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.555166 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2pv5\" (UniqueName: \"kubernetes.io/projected/f63b0e62-e9e8-432e-9a1b-f5be926fc2ce-kube-api-access-k2pv5\") pod \"node-resolver-r7x8v\" (UID: \"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\") " pod="openshift-dns/node-resolver-r7x8v" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.580118 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.603093 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.623053 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.641350 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.651766 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-r7x8v" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.674553 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.679507 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.684251 4857 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 01 21:35:01 crc kubenswrapper[4857]: W1201 21:35:01.685391 4857 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Dec 01 21:35:01 crc kubenswrapper[4857]: W1201 21:35:01.685539 4857 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Dec 01 21:35:01 crc kubenswrapper[4857]: W1201 21:35:01.685575 4857 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 01 21:35:01 crc kubenswrapper[4857]: W1201 21:35:01.685586 4857 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Dec 01 21:35:01 crc kubenswrapper[4857]: W1201 21:35:01.685595 4857 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 01 21:35:01 crc kubenswrapper[4857]: W1201 21:35:01.685602 4857 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 01 21:35:01 crc kubenswrapper[4857]: W1201 21:35:01.685622 4857 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Dec 01 21:35:01 crc kubenswrapper[4857]: W1201 21:35:01.687172 4857 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 01 21:35:01 crc kubenswrapper[4857]: W1201 21:35:01.687265 4857 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 01 21:35:01 crc kubenswrapper[4857]: W1201 21:35:01.697273 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac5dfac0_8933_4db8_9cb6_5dca0537d066.slice/crio-f99f8f49ae674c52fddc98caeb769967824a2270ee8f2f84824d6002b2cc5047 WatchSource:0}: Error finding container f99f8f49ae674c52fddc98caeb769967824a2270ee8f2f84824d6002b2cc5047: Status 404 returned error can't find the container with id f99f8f49ae674c52fddc98caeb769967824a2270ee8f2f84824d6002b2cc5047 Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.775605 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.775789 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:35:03.775767622 +0000 UTC m=+22.265829949 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.834006 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.834146 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.844900 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.845664 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.847270 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.848010 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.849199 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.849849 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.851234 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.852506 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.853484 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.854756 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.855410 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.856732 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.857387 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.857990 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.859179 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.859836 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.864468 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.864979 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.865775 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.866804 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.867369 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.867911 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.868727 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.869392 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.870234 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.870837 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.871858 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.872323 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.873264 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.873717 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.874190 4857 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.874669 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.876366 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.876830 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.876877 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.876894 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.876922 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.876947 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.877107 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.877175 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:03.877156332 +0000 UTC m=+22.367218649 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.877214 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.877237 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.877283 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.877298 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.877301 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:03.877279275 +0000 UTC m=+22.367341592 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.877342 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:03.877326346 +0000 UTC m=+22.367388663 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.877386 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.877398 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.877409 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:01 crc kubenswrapper[4857]: E1201 21:35:01.877433 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:03.877426769 +0000 UTC m=+22.367489086 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.877739 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.879215 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.880008 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.881059 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.882214 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.882845 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.883712 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.888422 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.889553 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.890887 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.891830 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.892576 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.893896 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.894814 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.895714 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.896385 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.897549 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.898545 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.899312 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.900459 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.900996 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-tczkh"] Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.901683 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mdc5b"] Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.902011 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.904092 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.905314 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-dfxtl"] Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.905362 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.905522 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.905619 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.906929 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.907355 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.907410 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.907590 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.907796 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.907820 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.908091 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.908155 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.908261 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.908528 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.908707 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.908860 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.953033 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85"} Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.953095 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786"} Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.953108 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"f99f8f49ae674c52fddc98caeb769967824a2270ee8f2f84824d6002b2cc5047"} Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.962663 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-r7x8v" event={"ID":"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce","Type":"ContainerStarted","Data":"6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f"} Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.962709 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-r7x8v" event={"ID":"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce","Type":"ContainerStarted","Data":"9a84832c7b549394cc008bc37d8f18c97385c904ad17ad5231abc35d849e0515"} Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.977980 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4m47\" (UniqueName: \"kubernetes.io/projected/6721c7e7-7840-4241-a3b2-d988c6dcf215-kube-api-access-f4m47\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.978265 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-multus-cni-dir\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.978390 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-systemd\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.978478 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-log-socket\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.978549 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6721c7e7-7840-4241-a3b2-d988c6dcf215-cnibin\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.978623 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-run-k8s-cni-cncf-io\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.978695 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-var-lib-cni-multus\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.978759 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-node-log\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.978848 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fc7d55e5-8085-40b7-8147-8508ee6c3f61-cni-binary-copy\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.978953 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-run-netns\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.979022 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-etc-kubernetes\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.979137 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-etc-openvswitch\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.979208 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovnkube-config\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.979290 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6721c7e7-7840-4241-a3b2-d988c6dcf215-tuning-conf-dir\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.979370 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-cnibin\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.979438 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-var-lib-cni-bin\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.979514 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-ovn\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.979597 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-cni-bin\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.979693 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6721c7e7-7840-4241-a3b2-d988c6dcf215-os-release\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.979791 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-openvswitch\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.979888 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovnkube-script-lib\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.980001 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-multus-conf-dir\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.980117 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovn-node-metrics-cert\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.980195 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-os-release\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.980273 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-var-lib-kubelet\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.980349 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-env-overrides\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.980440 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-kubelet\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.980541 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-run-netns\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.980634 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-cni-netd\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.980749 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-hostroot\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.980826 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-system-cni-dir\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.980921 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-run-multus-certs\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.981074 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97tnv\" (UniqueName: \"kubernetes.io/projected/fc7d55e5-8085-40b7-8147-8508ee6c3f61-kube-api-access-97tnv\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.981121 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftp2p\" (UniqueName: \"kubernetes.io/projected/557fd3cd-9ab9-4135-af14-f748fdc34a90-kube-api-access-ftp2p\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.981165 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6721c7e7-7840-4241-a3b2-d988c6dcf215-system-cni-dir\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.981195 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-slash\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.981230 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/fc7d55e5-8085-40b7-8147-8508ee6c3f61-multus-daemon-config\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.981261 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6721c7e7-7840-4241-a3b2-d988c6dcf215-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.981298 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6721c7e7-7840-4241-a3b2-d988c6dcf215-cni-binary-copy\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.981318 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-multus-socket-dir-parent\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.981335 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-run-ovn-kubernetes\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.981355 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-systemd-units\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.981396 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-var-lib-openvswitch\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:01 crc kubenswrapper[4857]: I1201 21:35:01.981414 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.082753 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6721c7e7-7840-4241-a3b2-d988c6dcf215-cni-binary-copy\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.082813 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-multus-socket-dir-parent\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.082859 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-run-ovn-kubernetes\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.082880 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-systemd-units\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.082917 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-var-lib-openvswitch\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.082939 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.082973 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4m47\" (UniqueName: \"kubernetes.io/projected/6721c7e7-7840-4241-a3b2-d988c6dcf215-kube-api-access-f4m47\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083012 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-multus-cni-dir\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083031 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-systemd\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083057 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-multus-socket-dir-parent\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083072 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-node-log\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083133 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-log-socket\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083139 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-node-log\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083155 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6721c7e7-7840-4241-a3b2-d988c6dcf215-cnibin\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083176 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6721c7e7-7840-4241-a3b2-d988c6dcf215-cnibin\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083196 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-log-socket\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083203 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-run-k8s-cni-cncf-io\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083227 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-var-lib-cni-multus\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083246 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fc7d55e5-8085-40b7-8147-8508ee6c3f61-cni-binary-copy\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083288 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-run-netns\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083307 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-etc-kubernetes\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083328 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-etc-openvswitch\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083369 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-cni-bin\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083389 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovnkube-config\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083438 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6721c7e7-7840-4241-a3b2-d988c6dcf215-tuning-conf-dir\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083462 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-cnibin\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083481 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-var-lib-cni-bin\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083500 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-var-lib-openvswitch\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083520 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-ovn\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083526 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083557 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6721c7e7-7840-4241-a3b2-d988c6dcf215-os-release\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.082982 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-run-ovn-kubernetes\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083599 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-openvswitch\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083622 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovn-node-metrics-cert\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083641 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovnkube-script-lib\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083705 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-multus-conf-dir\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083727 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-env-overrides\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083729 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-multus-cni-dir\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083765 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-ovn\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083769 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-os-release\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083790 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-cni-bin\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083787 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6721c7e7-7840-4241-a3b2-d988c6dcf215-cni-binary-copy\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083792 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-var-lib-kubelet\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083840 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-var-lib-kubelet\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083862 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-kubelet\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083880 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-hostroot\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083898 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-run-netns\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083913 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-cni-netd\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083943 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-system-cni-dir\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083960 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-run-multus-certs\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083976 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97tnv\" (UniqueName: \"kubernetes.io/projected/fc7d55e5-8085-40b7-8147-8508ee6c3f61-kube-api-access-97tnv\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083994 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftp2p\" (UniqueName: \"kubernetes.io/projected/557fd3cd-9ab9-4135-af14-f748fdc34a90-kube-api-access-ftp2p\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084024 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6721c7e7-7840-4241-a3b2-d988c6dcf215-system-cni-dir\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084059 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-slash\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084079 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6721c7e7-7840-4241-a3b2-d988c6dcf215-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084095 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/fc7d55e5-8085-40b7-8147-8508ee6c3f61-multus-daemon-config\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084447 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovnkube-config\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083560 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-etc-openvswitch\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084696 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6721c7e7-7840-4241-a3b2-d988c6dcf215-os-release\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084734 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-kubelet\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084764 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-hostroot\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084780 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-var-lib-cni-multus\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084804 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-cni-netd\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084832 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-cnibin\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084837 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-system-cni-dir\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084855 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-systemd\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.083484 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-systemd-units\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084846 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-openvswitch\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084787 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-run-netns\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084917 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-run-multus-certs\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.084925 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-etc-kubernetes\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.085271 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-multus-conf-dir\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.085488 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fc7d55e5-8085-40b7-8147-8508ee6c3f61-cni-binary-copy\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.085545 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-os-release\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.085565 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-run-k8s-cni-cncf-io\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.085618 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6721c7e7-7840-4241-a3b2-d988c6dcf215-tuning-conf-dir\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.085721 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6721c7e7-7840-4241-a3b2-d988c6dcf215-system-cni-dir\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.085729 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-run-netns\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.085741 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fc7d55e5-8085-40b7-8147-8508ee6c3f61-host-var-lib-cni-bin\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.085753 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-slash\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.085796 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovnkube-script-lib\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.085847 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-env-overrides\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.085855 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/fc7d55e5-8085-40b7-8147-8508ee6c3f61-multus-daemon-config\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.086164 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6721c7e7-7840-4241-a3b2-d988c6dcf215-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.091708 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovn-node-metrics-cert\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.106657 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97tnv\" (UniqueName: \"kubernetes.io/projected/fc7d55e5-8085-40b7-8147-8508ee6c3f61-kube-api-access-97tnv\") pod \"multus-dfxtl\" (UID: \"fc7d55e5-8085-40b7-8147-8508ee6c3f61\") " pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.108241 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftp2p\" (UniqueName: \"kubernetes.io/projected/557fd3cd-9ab9-4135-af14-f748fdc34a90-kube-api-access-ftp2p\") pod \"ovnkube-node-mdc5b\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.120705 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4m47\" (UniqueName: \"kubernetes.io/projected/6721c7e7-7840-4241-a3b2-d988c6dcf215-kube-api-access-f4m47\") pod \"multus-additional-cni-plugins-tczkh\" (UID: \"6721c7e7-7840-4241-a3b2-d988c6dcf215\") " pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.206890 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.212002 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.216749 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-tczkh" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.217176 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.227256 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:02 crc kubenswrapper[4857]: W1201 21:35:02.238155 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6721c7e7_7840_4241_a3b2_d988c6dcf215.slice/crio-bf878328cff98039a2464f7f12399d390faec74447d5dc75f4a9df0222d8fc42 WatchSource:0}: Error finding container bf878328cff98039a2464f7f12399d390faec74447d5dc75f4a9df0222d8fc42: Status 404 returned error can't find the container with id bf878328cff98039a2464f7f12399d390faec74447d5dc75f4a9df0222d8fc42 Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.239163 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-dfxtl" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.548114 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.698745 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.710813 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.718834 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.729319 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.743071 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.758360 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.760917 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.773875 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.788797 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.800327 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.802325 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.814919 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.829788 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.834016 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.834242 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:02 crc kubenswrapper[4857]: E1201 21:35:02.834420 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:02 crc kubenswrapper[4857]: E1201 21:35:02.834640 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.847813 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.865386 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.894782 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.914467 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.942395 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.957776 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:02 crc kubenswrapper[4857]: I1201 21:35:02.967015 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" event={"ID":"6721c7e7-7840-4241-a3b2-d988c6dcf215","Type":"ContainerStarted","Data":"bf878328cff98039a2464f7f12399d390faec74447d5dc75f4a9df0222d8fc42"} Dec 01 21:35:03 crc kubenswrapper[4857]: W1201 21:35:03.053425 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc7d55e5_8085_40b7_8147_8508ee6c3f61.slice/crio-5f8b434d062e5169a8ec74da6d2c13c44a8418202f2274c1c74c1f00ccdd041a WatchSource:0}: Error finding container 5f8b434d062e5169a8ec74da6d2c13c44a8418202f2274c1c74c1f00ccdd041a: Status 404 returned error can't find the container with id 5f8b434d062e5169a8ec74da6d2c13c44a8418202f2274c1c74c1f00ccdd041a Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.058427 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.058761 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.087687 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.171228 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.171228 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.171397 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.189822 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.223245 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.264489 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.265436 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.288132 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.302866 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.332685 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.371577 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.392499 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.478032 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-49cdz"] Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.478449 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-49cdz" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.480920 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.481350 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.482228 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.483850 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.503234 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.522278 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.552438 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.570159 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.593417 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.604021 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/112193dd-6798-4528-890e-a9713a21b3cb-serviceca\") pod \"node-ca-49cdz\" (UID: \"112193dd-6798-4528-890e-a9713a21b3cb\") " pod="openshift-image-registry/node-ca-49cdz" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.604095 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/112193dd-6798-4528-890e-a9713a21b3cb-host\") pod \"node-ca-49cdz\" (UID: \"112193dd-6798-4528-890e-a9713a21b3cb\") " pod="openshift-image-registry/node-ca-49cdz" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.604146 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g49r\" (UniqueName: \"kubernetes.io/projected/112193dd-6798-4528-890e-a9713a21b3cb-kube-api-access-4g49r\") pod \"node-ca-49cdz\" (UID: \"112193dd-6798-4528-890e-a9713a21b3cb\") " pod="openshift-image-registry/node-ca-49cdz" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.606978 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.619591 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.636611 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.658200 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.681365 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.690164 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.702528 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.704677 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g49r\" (UniqueName: \"kubernetes.io/projected/112193dd-6798-4528-890e-a9713a21b3cb-kube-api-access-4g49r\") pod \"node-ca-49cdz\" (UID: \"112193dd-6798-4528-890e-a9713a21b3cb\") " pod="openshift-image-registry/node-ca-49cdz" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.704730 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/112193dd-6798-4528-890e-a9713a21b3cb-serviceca\") pod \"node-ca-49cdz\" (UID: \"112193dd-6798-4528-890e-a9713a21b3cb\") " pod="openshift-image-registry/node-ca-49cdz" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.704751 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/112193dd-6798-4528-890e-a9713a21b3cb-host\") pod \"node-ca-49cdz\" (UID: \"112193dd-6798-4528-890e-a9713a21b3cb\") " pod="openshift-image-registry/node-ca-49cdz" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.704804 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/112193dd-6798-4528-890e-a9713a21b3cb-host\") pod \"node-ca-49cdz\" (UID: \"112193dd-6798-4528-890e-a9713a21b3cb\") " pod="openshift-image-registry/node-ca-49cdz" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.706415 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/112193dd-6798-4528-890e-a9713a21b3cb-serviceca\") pod \"node-ca-49cdz\" (UID: \"112193dd-6798-4528-890e-a9713a21b3cb\") " pod="openshift-image-registry/node-ca-49cdz" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.714337 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.722244 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g49r\" (UniqueName: \"kubernetes.io/projected/112193dd-6798-4528-890e-a9713a21b3cb-kube-api-access-4g49r\") pod \"node-ca-49cdz\" (UID: \"112193dd-6798-4528-890e-a9713a21b3cb\") " pod="openshift-image-registry/node-ca-49cdz" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.725613 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.805484 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.805732 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:35:07.805706566 +0000 UTC m=+26.295768883 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.846888 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.846981 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.883093 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-49cdz" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.915705 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.915749 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.915773 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.915800 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.915901 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.915953 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:07.915932334 +0000 UTC m=+26.405994651 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.916018 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.916028 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.916056 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.916083 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:07.916072558 +0000 UTC m=+26.406134875 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.916211 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.916250 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:07.916237322 +0000 UTC m=+26.406299639 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.916312 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.916348 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.916364 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:03 crc kubenswrapper[4857]: E1201 21:35:03.916446 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:07.916414257 +0000 UTC m=+26.406476574 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.970023 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-49cdz" event={"ID":"112193dd-6798-4528-890e-a9713a21b3cb","Type":"ContainerStarted","Data":"bae7c0f555fdd3c3de283fb585f56dfb4e94711aad86d0fede0b0d0fd534d4d9"} Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.971551 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f" exitCode=0 Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.971594 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f"} Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.971611 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"b7de8baf745835264f416445383bebf888433f4ef55305b0795912a2e4619616"} Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.983305 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" event={"ID":"6721c7e7-7840-4241-a3b2-d988c6dcf215","Type":"ContainerStarted","Data":"4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703"} Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.987074 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dfxtl" event={"ID":"fc7d55e5-8085-40b7-8147-8508ee6c3f61","Type":"ContainerStarted","Data":"8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3"} Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.987110 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dfxtl" event={"ID":"fc7d55e5-8085-40b7-8147-8508ee6c3f61","Type":"ContainerStarted","Data":"5f8b434d062e5169a8ec74da6d2c13c44a8418202f2274c1c74c1f00ccdd041a"} Dec 01 21:35:03 crc kubenswrapper[4857]: I1201 21:35:03.997409 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.009064 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.022557 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.044064 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.059546 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.076755 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.089450 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.099682 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.127728 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.139448 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.270663 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.286080 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.316693 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.326945 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.346165 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.357986 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.375968 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.494310 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.507106 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.519177 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.532219 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.544755 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.557706 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.611149 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.628633 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.628752 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.630644 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.630671 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.630680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.630810 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.637574 4857 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.637900 4857 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.638896 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.638929 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.638939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.638954 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.638964 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:04Z","lastTransitionTime":"2025-12-01T21:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.644535 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.656083 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: E1201 21:35:04.656988 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.660161 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.660197 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.660211 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.660227 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.660237 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:04Z","lastTransitionTime":"2025-12-01T21:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.670150 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: E1201 21:35:04.673811 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.676704 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.676737 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.676748 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.676764 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.676775 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:04Z","lastTransitionTime":"2025-12-01T21:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:04 crc kubenswrapper[4857]: E1201 21:35:04.686264 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.688857 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.688882 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.688890 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.688902 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.688911 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:04Z","lastTransitionTime":"2025-12-01T21:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:04 crc kubenswrapper[4857]: E1201 21:35:04.698530 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.700940 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.700961 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.700968 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.700979 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.700988 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:04Z","lastTransitionTime":"2025-12-01T21:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:04 crc kubenswrapper[4857]: E1201 21:35:04.711393 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:04 crc kubenswrapper[4857]: E1201 21:35:04.711541 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.712684 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.712713 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.712722 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.712733 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.712742 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:04Z","lastTransitionTime":"2025-12-01T21:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.814945 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.814979 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.814989 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.815003 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.815013 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:04Z","lastTransitionTime":"2025-12-01T21:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.834204 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.834204 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:04 crc kubenswrapper[4857]: E1201 21:35:04.834316 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:04 crc kubenswrapper[4857]: E1201 21:35:04.834380 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.918486 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.918518 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.918528 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.918544 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.918555 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:04Z","lastTransitionTime":"2025-12-01T21:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.995317 4857 generic.go:334] "Generic (PLEG): container finished" podID="6721c7e7-7840-4241-a3b2-d988c6dcf215" containerID="4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703" exitCode=0 Dec 01 21:35:04 crc kubenswrapper[4857]: I1201 21:35:04.995415 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" event={"ID":"6721c7e7-7840-4241-a3b2-d988c6dcf215","Type":"ContainerDied","Data":"4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:04.997655 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-49cdz" event={"ID":"112193dd-6798-4528-890e-a9713a21b3cb","Type":"ContainerStarted","Data":"69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.002641 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.002696 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.021106 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.021149 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.021162 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.021178 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.021190 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:05Z","lastTransitionTime":"2025-12-01T21:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.046690 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.061072 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.074197 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.097251 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.115586 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.130716 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.142990 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.159617 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.161388 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.161428 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.161441 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.161458 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.161469 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:05Z","lastTransitionTime":"2025-12-01T21:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.179586 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.215712 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.228248 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.267715 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.269317 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.269341 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.269350 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.269364 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.269373 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:05Z","lastTransitionTime":"2025-12-01T21:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.278562 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.292467 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.306122 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.323126 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.337223 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.356237 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.368678 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.371626 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.371648 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.371657 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.371670 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.371678 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:05Z","lastTransitionTime":"2025-12-01T21:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.379709 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.393272 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.408853 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.419515 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.430302 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.447110 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.458282 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.466958 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.473329 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.473357 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.473367 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.473382 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.473394 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:05Z","lastTransitionTime":"2025-12-01T21:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.479508 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.575688 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.575741 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.575759 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.575780 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.575802 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:05Z","lastTransitionTime":"2025-12-01T21:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.678280 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.678317 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.678331 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.678348 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.678359 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:05Z","lastTransitionTime":"2025-12-01T21:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.779983 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.780014 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.780022 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.780035 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.780058 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:05Z","lastTransitionTime":"2025-12-01T21:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.834937 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:05 crc kubenswrapper[4857]: E1201 21:35:05.835461 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.882758 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.882817 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.882840 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.882867 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.882887 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:05Z","lastTransitionTime":"2025-12-01T21:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.987083 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.987158 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.987173 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.987198 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:05 crc kubenswrapper[4857]: I1201 21:35:05.987220 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:05Z","lastTransitionTime":"2025-12-01T21:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.012380 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" event={"ID":"6721c7e7-7840-4241-a3b2-d988c6dcf215","Type":"ContainerStarted","Data":"7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.019349 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.019388 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.019401 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.019415 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.036887 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.043307 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.064240 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.080131 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.090152 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.090235 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.090247 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.090271 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.090287 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:06Z","lastTransitionTime":"2025-12-01T21:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.101786 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.119834 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.137918 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.157342 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.173364 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.190928 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.193012 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.193077 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.193094 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.193117 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.193136 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:06Z","lastTransitionTime":"2025-12-01T21:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.209217 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.227389 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.243275 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.261759 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.271601 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.283367 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.295743 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.295931 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.296022 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.296118 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.296184 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:06Z","lastTransitionTime":"2025-12-01T21:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.300606 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.323238 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.338130 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.350501 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.365399 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.392767 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.399177 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.399223 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.399238 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.399258 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.399273 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:06Z","lastTransitionTime":"2025-12-01T21:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.425284 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.452769 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.466822 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.478709 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.494671 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.502096 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.502150 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.502163 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.502182 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.502196 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:06Z","lastTransitionTime":"2025-12-01T21:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.510488 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.526309 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.604607 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.604647 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.604680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.604697 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.604709 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:06Z","lastTransitionTime":"2025-12-01T21:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.707449 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.707490 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.707499 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.707516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.707524 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:06Z","lastTransitionTime":"2025-12-01T21:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.810436 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.810501 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.810518 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.810544 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.810561 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:06Z","lastTransitionTime":"2025-12-01T21:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.834632 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.834675 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:06 crc kubenswrapper[4857]: E1201 21:35:06.834757 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:06 crc kubenswrapper[4857]: E1201 21:35:06.834870 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.912674 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.912731 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.912749 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.912771 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:06 crc kubenswrapper[4857]: I1201 21:35:06.912789 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:06Z","lastTransitionTime":"2025-12-01T21:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.020428 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.020487 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.020500 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.020516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.020525 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:07Z","lastTransitionTime":"2025-12-01T21:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.043093 4857 generic.go:334] "Generic (PLEG): container finished" podID="6721c7e7-7840-4241-a3b2-d988c6dcf215" containerID="7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec" exitCode=0 Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.043155 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" event={"ID":"6721c7e7-7840-4241-a3b2-d988c6dcf215","Type":"ContainerDied","Data":"7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec"} Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.068389 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.090360 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.103001 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.120357 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.123635 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.123668 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.123678 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.123695 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.123707 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:07Z","lastTransitionTime":"2025-12-01T21:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.133811 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.162252 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.182782 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.199467 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.216981 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.229718 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.229760 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.229778 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.229802 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.229820 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:07Z","lastTransitionTime":"2025-12-01T21:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.240931 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.254879 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.268708 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.289463 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.303774 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.331686 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.331728 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.331740 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.331757 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.331769 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:07Z","lastTransitionTime":"2025-12-01T21:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.434930 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.435012 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.435036 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.435098 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.435120 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:07Z","lastTransitionTime":"2025-12-01T21:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.538260 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.538302 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.538314 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.538330 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.538341 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:07Z","lastTransitionTime":"2025-12-01T21:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.641086 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.641416 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.641505 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.641526 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.641546 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:07Z","lastTransitionTime":"2025-12-01T21:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.745711 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.745770 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.745789 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.745820 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.745840 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:07Z","lastTransitionTime":"2025-12-01T21:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.834110 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:07 crc kubenswrapper[4857]: E1201 21:35:07.834282 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.848621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.848657 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.848670 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.848687 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.848699 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:07Z","lastTransitionTime":"2025-12-01T21:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.902365 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:07 crc kubenswrapper[4857]: E1201 21:35:07.902530 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:35:15.902500337 +0000 UTC m=+34.392562664 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.951175 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.951218 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.951231 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.951250 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:07 crc kubenswrapper[4857]: I1201 21:35:07.951264 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:07Z","lastTransitionTime":"2025-12-01T21:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.003542 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.003585 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.003603 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.003624 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.003686 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.003708 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.003752 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:16.003718652 +0000 UTC m=+34.493780969 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.003810 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:16.003790684 +0000 UTC m=+34.493853011 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.003846 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.003871 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.003886 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.003892 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.003956 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.003978 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.003933 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:16.003921558 +0000 UTC m=+34.493983955 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.004082 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:16.004031721 +0000 UTC m=+34.494094078 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.049719 4857 generic.go:334] "Generic (PLEG): container finished" podID="6721c7e7-7840-4241-a3b2-d988c6dcf215" containerID="0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc" exitCode=0 Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.049828 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" event={"ID":"6721c7e7-7840-4241-a3b2-d988c6dcf215","Type":"ContainerDied","Data":"0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc"} Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.052662 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.052710 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.052771 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.052804 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.052817 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:08Z","lastTransitionTime":"2025-12-01T21:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.057168 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c"} Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.073076 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.103927 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.131843 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.146220 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.155349 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.155383 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.155394 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.155409 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.155420 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:08Z","lastTransitionTime":"2025-12-01T21:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.160297 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.179514 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.193270 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.208133 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.219034 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.234494 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.253702 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.262117 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.262147 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.262156 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.262169 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.262179 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:08Z","lastTransitionTime":"2025-12-01T21:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.272152 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.285225 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.307183 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.364961 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.365003 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.365015 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.365034 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.365073 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:08Z","lastTransitionTime":"2025-12-01T21:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.467554 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.467612 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.467628 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.467650 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.467669 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:08Z","lastTransitionTime":"2025-12-01T21:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.570748 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.570789 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.570799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.570817 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.570827 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:08Z","lastTransitionTime":"2025-12-01T21:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.673862 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.673895 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.673903 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.673915 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.673925 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:08Z","lastTransitionTime":"2025-12-01T21:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.776913 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.776971 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.776988 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.777019 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.777064 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:08Z","lastTransitionTime":"2025-12-01T21:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.834246 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.834246 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.834450 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:08 crc kubenswrapper[4857]: E1201 21:35:08.834667 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.879851 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.879892 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.879903 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.879921 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.879936 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:08Z","lastTransitionTime":"2025-12-01T21:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.982237 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.982284 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.982294 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.982310 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:08 crc kubenswrapper[4857]: I1201 21:35:08.982321 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:08Z","lastTransitionTime":"2025-12-01T21:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.064799 4857 generic.go:334] "Generic (PLEG): container finished" podID="6721c7e7-7840-4241-a3b2-d988c6dcf215" containerID="492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca" exitCode=0 Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.064850 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" event={"ID":"6721c7e7-7840-4241-a3b2-d988c6dcf215","Type":"ContainerDied","Data":"492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca"} Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.084855 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.084890 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.084899 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.084916 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.084925 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:09Z","lastTransitionTime":"2025-12-01T21:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.090972 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.107740 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.123897 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.142700 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.158325 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.180033 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.188674 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.188754 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.188774 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.188800 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.188821 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:09Z","lastTransitionTime":"2025-12-01T21:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.197875 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.218488 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.237977 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.259955 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.280556 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.292914 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.292954 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.292968 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.292988 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.293004 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:09Z","lastTransitionTime":"2025-12-01T21:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.296353 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.321882 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.335129 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:09Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.395867 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.395898 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.395908 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.395923 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.395933 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:09Z","lastTransitionTime":"2025-12-01T21:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.500659 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.500712 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.500723 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.500741 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.500752 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:09Z","lastTransitionTime":"2025-12-01T21:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.603215 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.603276 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.603293 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.603316 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.603336 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:09Z","lastTransitionTime":"2025-12-01T21:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.706751 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.706800 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.706816 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.706840 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.706866 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:09Z","lastTransitionTime":"2025-12-01T21:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.809584 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.809653 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.809675 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.809702 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.809720 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:09Z","lastTransitionTime":"2025-12-01T21:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.834817 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:09 crc kubenswrapper[4857]: E1201 21:35:09.834988 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.916956 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.917016 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.917070 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.917114 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:09 crc kubenswrapper[4857]: I1201 21:35:09.917137 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:09Z","lastTransitionTime":"2025-12-01T21:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.019913 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.020215 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.020234 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.020253 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.020265 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:10Z","lastTransitionTime":"2025-12-01T21:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.073431 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" event={"ID":"6721c7e7-7840-4241-a3b2-d988c6dcf215","Type":"ContainerStarted","Data":"79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f"} Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.092456 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.106594 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.122052 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.122122 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.122152 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.122161 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.122175 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.122185 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:10Z","lastTransitionTime":"2025-12-01T21:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.137950 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.153820 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.175231 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.192313 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.210112 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.225345 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.225582 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.225591 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.225618 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.225631 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:10Z","lastTransitionTime":"2025-12-01T21:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.225691 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.243120 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.262485 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.276231 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.288691 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.304382 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:10Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.327830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.327874 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.327886 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.327901 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.327912 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:10Z","lastTransitionTime":"2025-12-01T21:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.430176 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.430225 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.430243 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.430269 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.430285 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:10Z","lastTransitionTime":"2025-12-01T21:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.532796 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.532838 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.532848 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.532864 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.532875 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:10Z","lastTransitionTime":"2025-12-01T21:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.635414 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.635449 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.635458 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.635473 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.635483 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:10Z","lastTransitionTime":"2025-12-01T21:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.738202 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.738368 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.738505 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.738665 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.738734 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:10Z","lastTransitionTime":"2025-12-01T21:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.834791 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.834850 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:10 crc kubenswrapper[4857]: E1201 21:35:10.834974 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:10 crc kubenswrapper[4857]: E1201 21:35:10.835078 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.841328 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.841363 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.841372 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.841395 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.841406 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:10Z","lastTransitionTime":"2025-12-01T21:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.926882 4857 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.943840 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.943888 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.943903 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.943922 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:10 crc kubenswrapper[4857]: I1201 21:35:10.943934 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:10Z","lastTransitionTime":"2025-12-01T21:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.046947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.047006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.047026 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.047078 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.047097 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:11Z","lastTransitionTime":"2025-12-01T21:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.079573 4857 generic.go:334] "Generic (PLEG): container finished" podID="6721c7e7-7840-4241-a3b2-d988c6dcf215" containerID="79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f" exitCode=0 Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.079653 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" event={"ID":"6721c7e7-7840-4241-a3b2-d988c6dcf215","Type":"ContainerDied","Data":"79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f"} Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.086111 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591"} Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.086606 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.086676 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.098417 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.112263 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.126633 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.142918 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.147401 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.149839 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.150175 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.150256 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.150277 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.150299 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.150355 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:11Z","lastTransitionTime":"2025-12-01T21:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.158883 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.246092 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.254980 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.255019 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.255031 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.255071 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.255084 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:11Z","lastTransitionTime":"2025-12-01T21:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.267239 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.286018 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.303073 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.326396 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.341389 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.358585 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.358623 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.358635 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.358651 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.358664 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:11Z","lastTransitionTime":"2025-12-01T21:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.359672 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.373304 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.385755 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.408740 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.420678 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.440500 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.454596 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.460410 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.460447 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.460458 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.460472 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.460486 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:11Z","lastTransitionTime":"2025-12-01T21:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.466985 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.483389 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.498532 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.512243 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.527152 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.542724 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.560753 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.599553 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.599608 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.599626 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.599651 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.599667 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:11Z","lastTransitionTime":"2025-12-01T21:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.602946 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.620467 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.661376 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.702557 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.702608 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.702620 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.702636 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.702645 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:11Z","lastTransitionTime":"2025-12-01T21:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.804675 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.804743 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.804754 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.804772 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.804785 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:11Z","lastTransitionTime":"2025-12-01T21:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.834196 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:11 crc kubenswrapper[4857]: E1201 21:35:11.834421 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.849832 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.871415 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.887475 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.901981 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.907839 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.908501 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.908593 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.908684 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.908769 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:11Z","lastTransitionTime":"2025-12-01T21:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.925544 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.949574 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.963687 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:11 crc kubenswrapper[4857]: I1201 21:35:11.987940 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.001597 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.011957 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.012022 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.012057 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.012088 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.012103 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:12Z","lastTransitionTime":"2025-12-01T21:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.016883 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.032585 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.047068 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.064220 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.077910 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.092680 4857 generic.go:334] "Generic (PLEG): container finished" podID="6721c7e7-7840-4241-a3b2-d988c6dcf215" containerID="c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31" exitCode=0 Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.092747 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" event={"ID":"6721c7e7-7840-4241-a3b2-d988c6dcf215","Type":"ContainerDied","Data":"c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31"} Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.092864 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.107297 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.117707 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.117754 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.117767 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.117798 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.117810 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:12Z","lastTransitionTime":"2025-12-01T21:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.120566 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.134442 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.151859 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.168331 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.188969 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.203725 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.217190 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.220088 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.220122 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.220131 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.220145 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.220154 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:12Z","lastTransitionTime":"2025-12-01T21:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.233973 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.262238 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.277562 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.295095 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.307866 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.318021 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.322937 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.322976 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.322990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.323011 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.323025 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:12Z","lastTransitionTime":"2025-12-01T21:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.425021 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.425093 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.425108 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.425127 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.425141 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:12Z","lastTransitionTime":"2025-12-01T21:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.527629 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.527665 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.527676 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.527689 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.527698 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:12Z","lastTransitionTime":"2025-12-01T21:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.630309 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.630359 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.630373 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.630392 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.630406 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:12Z","lastTransitionTime":"2025-12-01T21:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.733604 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.733665 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.733683 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.733709 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.733727 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:12Z","lastTransitionTime":"2025-12-01T21:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.834871 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.834930 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:12 crc kubenswrapper[4857]: E1201 21:35:12.834995 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:12 crc kubenswrapper[4857]: E1201 21:35:12.835084 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.837155 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.837214 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.837226 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.837237 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.837246 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:12Z","lastTransitionTime":"2025-12-01T21:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.940267 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.940296 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.940306 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.940322 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:12 crc kubenswrapper[4857]: I1201 21:35:12.940333 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:12Z","lastTransitionTime":"2025-12-01T21:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.043647 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.043695 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.043712 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.043737 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.043756 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:13Z","lastTransitionTime":"2025-12-01T21:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.103992 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.104151 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" event={"ID":"6721c7e7-7840-4241-a3b2-d988c6dcf215","Type":"ContainerStarted","Data":"94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989"} Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.129659 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.146905 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.146954 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.146966 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.146984 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.146997 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:13Z","lastTransitionTime":"2025-12-01T21:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.171406 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.192333 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.217763 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.244785 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.249091 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.249156 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.249176 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.249202 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.249222 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:13Z","lastTransitionTime":"2025-12-01T21:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.266227 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.285837 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.305191 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.345778 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.351325 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.351385 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.351397 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.351421 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.351435 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:13Z","lastTransitionTime":"2025-12-01T21:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.375300 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.388297 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.403933 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.421652 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.431966 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:13Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.453672 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.453717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.453733 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.453754 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.453767 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:13Z","lastTransitionTime":"2025-12-01T21:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.556085 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.556137 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.556149 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.556169 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.556181 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:13Z","lastTransitionTime":"2025-12-01T21:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.659197 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.659294 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.659310 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.659342 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.659360 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:13Z","lastTransitionTime":"2025-12-01T21:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.762590 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.762658 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.762677 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.762704 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.762721 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:13Z","lastTransitionTime":"2025-12-01T21:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.834808 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:13 crc kubenswrapper[4857]: E1201 21:35:13.835104 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.865404 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.865466 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.865481 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.865504 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.865520 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:13Z","lastTransitionTime":"2025-12-01T21:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.968318 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.968379 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.968393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.968416 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:13 crc kubenswrapper[4857]: I1201 21:35:13.968434 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:13Z","lastTransitionTime":"2025-12-01T21:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.072327 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.072374 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.072386 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.072405 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.072417 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:14Z","lastTransitionTime":"2025-12-01T21:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.175269 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.175322 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.175333 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.175355 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.175367 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:14Z","lastTransitionTime":"2025-12-01T21:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.278570 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.278625 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.278635 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.278654 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.278665 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:14Z","lastTransitionTime":"2025-12-01T21:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.381984 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.382111 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.382146 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.382183 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.382206 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:14Z","lastTransitionTime":"2025-12-01T21:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.486974 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.487086 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.487107 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.487136 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.487153 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:14Z","lastTransitionTime":"2025-12-01T21:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.591225 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.591303 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.591322 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.591353 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.591371 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:14Z","lastTransitionTime":"2025-12-01T21:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.694869 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.694933 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.694954 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.694982 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.695005 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:14Z","lastTransitionTime":"2025-12-01T21:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.753387 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr"] Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.754342 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.757513 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.758255 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.778440 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.780082 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3c8e9109-4f8d-4410-a463-368cb7cfbee9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-c8mcr\" (UID: \"3c8e9109-4f8d-4410-a463-368cb7cfbee9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.780170 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7498\" (UniqueName: \"kubernetes.io/projected/3c8e9109-4f8d-4410-a463-368cb7cfbee9-kube-api-access-c7498\") pod \"ovnkube-control-plane-749d76644c-c8mcr\" (UID: \"3c8e9109-4f8d-4410-a463-368cb7cfbee9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.780269 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3c8e9109-4f8d-4410-a463-368cb7cfbee9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-c8mcr\" (UID: \"3c8e9109-4f8d-4410-a463-368cb7cfbee9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.780471 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3c8e9109-4f8d-4410-a463-368cb7cfbee9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-c8mcr\" (UID: \"3c8e9109-4f8d-4410-a463-368cb7cfbee9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.798579 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.798641 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.798665 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.798696 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.798715 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:14Z","lastTransitionTime":"2025-12-01T21:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.805294 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.834288 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:14 crc kubenswrapper[4857]: E1201 21:35:14.834504 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.834304 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:14 crc kubenswrapper[4857]: E1201 21:35:14.835148 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.837028 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.864224 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.882141 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3c8e9109-4f8d-4410-a463-368cb7cfbee9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-c8mcr\" (UID: \"3c8e9109-4f8d-4410-a463-368cb7cfbee9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.882273 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3c8e9109-4f8d-4410-a463-368cb7cfbee9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-c8mcr\" (UID: \"3c8e9109-4f8d-4410-a463-368cb7cfbee9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.882338 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7498\" (UniqueName: \"kubernetes.io/projected/3c8e9109-4f8d-4410-a463-368cb7cfbee9-kube-api-access-c7498\") pod \"ovnkube-control-plane-749d76644c-c8mcr\" (UID: \"3c8e9109-4f8d-4410-a463-368cb7cfbee9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.882410 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3c8e9109-4f8d-4410-a463-368cb7cfbee9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-c8mcr\" (UID: \"3c8e9109-4f8d-4410-a463-368cb7cfbee9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.883335 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.883450 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3c8e9109-4f8d-4410-a463-368cb7cfbee9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-c8mcr\" (UID: \"3c8e9109-4f8d-4410-a463-368cb7cfbee9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.883924 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3c8e9109-4f8d-4410-a463-368cb7cfbee9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-c8mcr\" (UID: \"3c8e9109-4f8d-4410-a463-368cb7cfbee9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.901422 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.901493 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.901509 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.901540 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.901557 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:14Z","lastTransitionTime":"2025-12-01T21:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.902227 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.908879 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3c8e9109-4f8d-4410-a463-368cb7cfbee9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-c8mcr\" (UID: \"3c8e9109-4f8d-4410-a463-368cb7cfbee9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.917462 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7498\" (UniqueName: \"kubernetes.io/projected/3c8e9109-4f8d-4410-a463-368cb7cfbee9-kube-api-access-c7498\") pod \"ovnkube-control-plane-749d76644c-c8mcr\" (UID: \"3c8e9109-4f8d-4410-a463-368cb7cfbee9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.925911 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.951888 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:14 crc kubenswrapper[4857]: I1201 21:35:14.970261 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.054163 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.059372 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.059413 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.059423 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.059437 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.059448 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: E1201 21:35:15.070741 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.071271 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.075169 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.075223 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.075233 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.075255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.075266 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.077601 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.086787 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: E1201 21:35:15.087324 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.091273 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.091435 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.091729 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.091841 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.091949 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: W1201 21:35:15.094457 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c8e9109_4f8d_4410_a463_368cb7cfbee9.slice/crio-e0b9bd86c7400d490e0e7cdf44e805a1471f346191f96e313f2a5821c0f1ab56 WatchSource:0}: Error finding container e0b9bd86c7400d490e0e7cdf44e805a1471f346191f96e313f2a5821c0f1ab56: Status 404 returned error can't find the container with id e0b9bd86c7400d490e0e7cdf44e805a1471f346191f96e313f2a5821c0f1ab56 Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.101171 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: E1201 21:35:15.106702 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.112680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.114534 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" event={"ID":"3c8e9109-4f8d-4410-a463-368cb7cfbee9","Type":"ContainerStarted","Data":"e0b9bd86c7400d490e0e7cdf44e805a1471f346191f96e313f2a5821c0f1ab56"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.114550 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.114617 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.114639 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.114654 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.117235 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/0.log" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.118532 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.119961 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591" exitCode=1 Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.120017 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.121976 4857 scope.go:117] "RemoveContainer" containerID="941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591" Dec 01 21:35:15 crc kubenswrapper[4857]: E1201 21:35:15.130023 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.131006 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.134025 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.134064 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.134075 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.134088 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.134099 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.144312 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: E1201 21:35:15.145906 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: E1201 21:35:15.146018 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.148801 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.148830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.148839 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.148853 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.148873 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.161606 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.176982 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.193023 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.206683 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.220958 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.233888 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.245573 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.251155 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.251204 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.251216 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.251239 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.251254 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.255798 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.268551 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.280992 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.299616 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"g/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:14.116172 6072 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:14.116212 6072 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:14.116229 6072 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:14.116239 6072 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 21:35:14.116256 6072 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 21:35:14.116264 6072 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:14.116270 6072 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 21:35:14.116286 6072 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 21:35:14.116298 6072 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:14.116333 6072 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 21:35:14.116350 6072 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 21:35:14.116375 6072 factory.go:656] Stopping watch factory\\\\nI1201 21:35:14.116395 6072 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:14.116387 6072 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 21:35:14.116430 6072 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.316642 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.332673 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.352361 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.353927 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.353968 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.353981 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.353998 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.354011 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.457687 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.457800 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.457827 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.457859 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.457883 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.561641 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.561721 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.561745 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.561807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.561831 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.665225 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.665282 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.665298 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.665321 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.665338 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.767380 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.767445 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.767469 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.767500 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.767552 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.834375 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:15 crc kubenswrapper[4857]: E1201 21:35:15.834552 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.871681 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.871748 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.871773 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.871807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.871831 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.896494 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-n2fsz"] Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.897259 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:15 crc kubenswrapper[4857]: E1201 21:35:15.897347 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.929957 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.952019 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.968724 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.975452 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.975495 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.975515 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.975542 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.975561 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:15Z","lastTransitionTime":"2025-12-01T21:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.989565 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:15Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.995423 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.995546 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:15 crc kubenswrapper[4857]: E1201 21:35:15.995601 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:35:31.995576936 +0000 UTC m=+50.485639253 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:35:15 crc kubenswrapper[4857]: I1201 21:35:15.995637 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss5tk\" (UniqueName: \"kubernetes.io/projected/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-kube-api-access-ss5tk\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.008564 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.028136 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.043509 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.060901 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.076257 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.078275 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.078303 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.078313 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.078329 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.078338 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:16Z","lastTransitionTime":"2025-12-01T21:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.091795 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.096376 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss5tk\" (UniqueName: \"kubernetes.io/projected/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-kube-api-access-ss5tk\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.096412 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.096434 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.096461 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.096479 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.096498 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096565 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096605 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:32.096593812 +0000 UTC m=+50.586656129 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096656 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096673 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096666 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096729 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096782 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096686 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096803 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096686 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096794 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:32.096766106 +0000 UTC m=+50.586828423 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096893 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:32.096880219 +0000 UTC m=+50.586942536 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096914 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:32.096899429 +0000 UTC m=+50.586961746 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.096930 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs podName:485d0d95-8002-4bc3-8935-8b8f4ab5b2b3 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:16.59692266 +0000 UTC m=+35.086984977 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs") pod "network-metrics-daemon-n2fsz" (UID: "485d0d95-8002-4bc3-8935-8b8f4ab5b2b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.104215 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.116407 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss5tk\" (UniqueName: \"kubernetes.io/projected/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-kube-api-access-ss5tk\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.123645 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.130674 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/0.log" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.134347 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a"} Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.134485 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.136097 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" event={"ID":"3c8e9109-4f8d-4410-a463-368cb7cfbee9","Type":"ContainerStarted","Data":"01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f"} Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.141931 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.164894 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"g/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:14.116172 6072 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:14.116212 6072 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:14.116229 6072 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:14.116239 6072 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 21:35:14.116256 6072 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 21:35:14.116264 6072 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:14.116270 6072 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 21:35:14.116286 6072 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 21:35:14.116298 6072 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:14.116333 6072 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 21:35:14.116350 6072 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 21:35:14.116375 6072 factory.go:656] Stopping watch factory\\\\nI1201 21:35:14.116395 6072 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:14.116387 6072 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 21:35:14.116430 6072 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.179125 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.180624 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.180657 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.180669 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.180688 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.180699 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:16Z","lastTransitionTime":"2025-12-01T21:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.203584 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.226975 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.250334 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.270473 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.282278 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.283016 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.283066 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.283079 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.283095 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.283106 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:16Z","lastTransitionTime":"2025-12-01T21:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.294408 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.310068 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.331349 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"g/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:14.116172 6072 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:14.116212 6072 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:14.116229 6072 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:14.116239 6072 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 21:35:14.116256 6072 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 21:35:14.116264 6072 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:14.116270 6072 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 21:35:14.116286 6072 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 21:35:14.116298 6072 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:14.116333 6072 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 21:35:14.116350 6072 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 21:35:14.116375 6072 factory.go:656] Stopping watch factory\\\\nI1201 21:35:14.116395 6072 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:14.116387 6072 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 21:35:14.116430 6072 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.343562 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.355974 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.372288 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.385851 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.385896 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.385909 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.385927 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.385976 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:16Z","lastTransitionTime":"2025-12-01T21:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.396114 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.409094 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.424611 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.438758 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.451698 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.465712 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:16Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.492674 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.492721 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.492733 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.492751 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.492765 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:16Z","lastTransitionTime":"2025-12-01T21:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.595470 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.595526 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.595541 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.595562 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.595581 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:16Z","lastTransitionTime":"2025-12-01T21:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.602026 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.602160 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.602228 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs podName:485d0d95-8002-4bc3-8935-8b8f4ab5b2b3 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:17.602210604 +0000 UTC m=+36.092272941 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs") pod "network-metrics-daemon-n2fsz" (UID: "485d0d95-8002-4bc3-8935-8b8f4ab5b2b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.698196 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.698241 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.698264 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.698281 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.698310 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:16Z","lastTransitionTime":"2025-12-01T21:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.800266 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.800303 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.800314 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.800330 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.800342 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:16Z","lastTransitionTime":"2025-12-01T21:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.833828 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.833939 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.834283 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:16 crc kubenswrapper[4857]: E1201 21:35:16.834504 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.902568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.902599 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.902609 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.902630 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:16 crc kubenswrapper[4857]: I1201 21:35:16.902654 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:16Z","lastTransitionTime":"2025-12-01T21:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.006125 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.006182 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.006196 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.006215 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.006228 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:17Z","lastTransitionTime":"2025-12-01T21:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.109297 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.109363 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.109404 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.109437 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.109461 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:17Z","lastTransitionTime":"2025-12-01T21:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.141073 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" event={"ID":"3c8e9109-4f8d-4410-a463-368cb7cfbee9","Type":"ContainerStarted","Data":"f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d"} Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.144538 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/1.log" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.145496 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/0.log" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.148891 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a" exitCode=1 Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.148930 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a"} Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.148966 4857 scope.go:117] "RemoveContainer" containerID="941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.149785 4857 scope.go:117] "RemoveContainer" containerID="d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a" Dec 01 21:35:17 crc kubenswrapper[4857]: E1201 21:35:17.149979 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.164413 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.184135 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.197917 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.214514 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.214623 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.214796 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.214818 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.214841 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.214885 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:17Z","lastTransitionTime":"2025-12-01T21:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.230951 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.250562 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.268504 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.283898 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.298148 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.315858 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.336355 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.336412 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.336423 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.336445 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.336457 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:17Z","lastTransitionTime":"2025-12-01T21:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.336679 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.352548 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.367423 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.389412 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"g/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:14.116172 6072 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:14.116212 6072 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:14.116229 6072 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:14.116239 6072 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 21:35:14.116256 6072 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 21:35:14.116264 6072 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:14.116270 6072 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 21:35:14.116286 6072 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 21:35:14.116298 6072 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:14.116333 6072 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 21:35:14.116350 6072 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 21:35:14.116375 6072 factory.go:656] Stopping watch factory\\\\nI1201 21:35:14.116395 6072 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:14.116387 6072 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 21:35:14.116430 6072 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.404496 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.421381 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.438163 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.439596 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.439646 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.439659 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.439680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.439693 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:17Z","lastTransitionTime":"2025-12-01T21:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.455116 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.472484 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.494705 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.511146 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.530558 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.541933 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.541983 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.542002 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.542028 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.542076 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:17Z","lastTransitionTime":"2025-12-01T21:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.547109 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.568263 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.582805 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.601901 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.610584 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:17 crc kubenswrapper[4857]: E1201 21:35:17.610812 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:17 crc kubenswrapper[4857]: E1201 21:35:17.610915 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs podName:485d0d95-8002-4bc3-8935-8b8f4ab5b2b3 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:19.610893169 +0000 UTC m=+38.100955486 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs") pod "network-metrics-daemon-n2fsz" (UID: "485d0d95-8002-4bc3-8935-8b8f4ab5b2b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.631392 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://941f0008f1448c0700be198ce84b6b4dcace6eb232e777a3e2cca5551eafe591\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"message\\\":\\\"g/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:14.116172 6072 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:14.116212 6072 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:14.116229 6072 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:14.116239 6072 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 21:35:14.116256 6072 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 21:35:14.116264 6072 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:14.116270 6072 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 21:35:14.116286 6072 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 21:35:14.116298 6072 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:14.116333 6072 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 21:35:14.116350 6072 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 21:35:14.116375 6072 factory.go:656] Stopping watch factory\\\\nI1201 21:35:14.116395 6072 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:14.116387 6072 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 21:35:14.116430 6072 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:16Z\\\",\\\"message\\\":\\\"/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:16.788564 6247 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789027 6247 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789410 6247 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 21:35:16.790116 6247 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:16.790151 6247 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:16.790158 6247 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:16.790175 6247 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 21:35:16.790254 6247 factory.go:656] Stopping watch factory\\\\nI1201 21:35:16.790277 6247 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:16.790304 6247 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 21:35:16.790317 6247 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:16.790326 6247 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.645877 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.645964 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.646000 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.646101 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.646127 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:17Z","lastTransitionTime":"2025-12-01T21:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.649852 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.668378 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.685295 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.706549 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.725655 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.750496 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.750543 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.750557 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.750575 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.750588 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:17Z","lastTransitionTime":"2025-12-01T21:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.834167 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:17 crc kubenswrapper[4857]: E1201 21:35:17.834358 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.834167 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:17 crc kubenswrapper[4857]: E1201 21:35:17.834545 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.853959 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.854024 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.854071 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.854095 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.854110 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:17Z","lastTransitionTime":"2025-12-01T21:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.957072 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.957168 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.957180 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.957198 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:17 crc kubenswrapper[4857]: I1201 21:35:17.957209 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:17Z","lastTransitionTime":"2025-12-01T21:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.060663 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.060719 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.060736 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.060759 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.060778 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:18Z","lastTransitionTime":"2025-12-01T21:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.155671 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/1.log" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.163161 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.163214 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.163233 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.163257 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.163276 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:18Z","lastTransitionTime":"2025-12-01T21:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.266560 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.266680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.266702 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.266732 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.266752 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:18Z","lastTransitionTime":"2025-12-01T21:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.369773 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.369860 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.369877 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.369904 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.369923 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:18Z","lastTransitionTime":"2025-12-01T21:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.473376 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.473424 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.473476 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.473528 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.473549 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:18Z","lastTransitionTime":"2025-12-01T21:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.575607 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.575659 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.575675 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.575691 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.575705 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:18Z","lastTransitionTime":"2025-12-01T21:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.684896 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.685096 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.685162 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.685267 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.685293 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:18Z","lastTransitionTime":"2025-12-01T21:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.788156 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.788234 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.788253 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.788276 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.788292 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:18Z","lastTransitionTime":"2025-12-01T21:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.834704 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.834827 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:18 crc kubenswrapper[4857]: E1201 21:35:18.834914 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:18 crc kubenswrapper[4857]: E1201 21:35:18.835113 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.892454 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.892513 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.892528 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.892552 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.892569 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:18Z","lastTransitionTime":"2025-12-01T21:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.994940 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.995011 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.995036 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.995102 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:18 crc kubenswrapper[4857]: I1201 21:35:18.995131 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:18Z","lastTransitionTime":"2025-12-01T21:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.098444 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.098491 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.098502 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.098521 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.098532 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:19Z","lastTransitionTime":"2025-12-01T21:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.202506 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.202563 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.202580 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.202603 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.202622 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:19Z","lastTransitionTime":"2025-12-01T21:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.305930 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.305969 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.305977 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.305993 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.306003 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:19Z","lastTransitionTime":"2025-12-01T21:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.409227 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.409303 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.409314 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.409330 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.409342 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:19Z","lastTransitionTime":"2025-12-01T21:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.512136 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.512218 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.512242 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.512271 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.512293 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:19Z","lastTransitionTime":"2025-12-01T21:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.614339 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.614393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.614411 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.614433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.614450 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:19Z","lastTransitionTime":"2025-12-01T21:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.635595 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:19 crc kubenswrapper[4857]: E1201 21:35:19.635769 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:19 crc kubenswrapper[4857]: E1201 21:35:19.635841 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs podName:485d0d95-8002-4bc3-8935-8b8f4ab5b2b3 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:23.635819416 +0000 UTC m=+42.125881763 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs") pod "network-metrics-daemon-n2fsz" (UID: "485d0d95-8002-4bc3-8935-8b8f4ab5b2b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.717471 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.717556 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.717584 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.717616 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.717641 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:19Z","lastTransitionTime":"2025-12-01T21:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.820501 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.820568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.820587 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.820613 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.820632 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:19Z","lastTransitionTime":"2025-12-01T21:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.834646 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.834736 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:19 crc kubenswrapper[4857]: E1201 21:35:19.834831 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:19 crc kubenswrapper[4857]: E1201 21:35:19.835030 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.923355 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.923423 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.923447 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.923474 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:19 crc kubenswrapper[4857]: I1201 21:35:19.923495 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:19Z","lastTransitionTime":"2025-12-01T21:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.026693 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.026758 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.026783 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.026821 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.026844 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:20Z","lastTransitionTime":"2025-12-01T21:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.130068 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.130126 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.130145 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.130170 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.130189 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:20Z","lastTransitionTime":"2025-12-01T21:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.233026 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.233110 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.233126 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.233146 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.233163 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:20Z","lastTransitionTime":"2025-12-01T21:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.335426 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.335477 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.335491 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.335510 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.335523 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:20Z","lastTransitionTime":"2025-12-01T21:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.439144 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.439214 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.439233 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.439259 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.439276 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:20Z","lastTransitionTime":"2025-12-01T21:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.542356 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.542397 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.542406 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.542422 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.542431 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:20Z","lastTransitionTime":"2025-12-01T21:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.645538 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.645595 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.645604 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.645619 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.645632 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:20Z","lastTransitionTime":"2025-12-01T21:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.772740 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.772830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.772863 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.772896 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.772916 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:20Z","lastTransitionTime":"2025-12-01T21:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.833859 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:20 crc kubenswrapper[4857]: E1201 21:35:20.834269 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.833877 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:20 crc kubenswrapper[4857]: E1201 21:35:20.834519 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.875440 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.875530 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.875558 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.875594 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.875616 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:20Z","lastTransitionTime":"2025-12-01T21:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.914621 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.915464 4857 scope.go:117] "RemoveContainer" containerID="d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a" Dec 01 21:35:20 crc kubenswrapper[4857]: E1201 21:35:20.915667 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.931703 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.957076 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.975222 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.978112 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.978239 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.978330 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.978447 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.978531 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:20Z","lastTransitionTime":"2025-12-01T21:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:20 crc kubenswrapper[4857]: I1201 21:35:20.991461 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.005453 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.021110 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.039963 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.057082 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.072814 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.080777 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.080829 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.080850 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.080874 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.080891 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:21Z","lastTransitionTime":"2025-12-01T21:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.086569 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.101197 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.130702 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:16Z\\\",\\\"message\\\":\\\"/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:16.788564 6247 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789027 6247 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789410 6247 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 21:35:16.790116 6247 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:16.790151 6247 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:16.790158 6247 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:16.790175 6247 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 21:35:16.790254 6247 factory.go:656] Stopping watch factory\\\\nI1201 21:35:16.790277 6247 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:16.790304 6247 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 21:35:16.790317 6247 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:16.790326 6247 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.144419 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.161992 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.176554 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.183378 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.183494 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.183515 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.183541 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.183559 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:21Z","lastTransitionTime":"2025-12-01T21:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.194936 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.286699 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.286762 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.286785 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.286817 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.286840 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:21Z","lastTransitionTime":"2025-12-01T21:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.389542 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.389606 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.389628 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.389655 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.389673 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:21Z","lastTransitionTime":"2025-12-01T21:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.492730 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.492790 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.492803 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.492830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.492845 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:21Z","lastTransitionTime":"2025-12-01T21:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.596161 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.596222 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.596243 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.596271 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.596293 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:21Z","lastTransitionTime":"2025-12-01T21:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.699329 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.699396 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.699413 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.699436 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.699453 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:21Z","lastTransitionTime":"2025-12-01T21:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.803171 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.803215 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.803227 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.803244 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.803254 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:21Z","lastTransitionTime":"2025-12-01T21:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.834108 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.834165 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:21 crc kubenswrapper[4857]: E1201 21:35:21.834288 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:21 crc kubenswrapper[4857]: E1201 21:35:21.834617 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.865753 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.885663 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.906707 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.906774 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.906787 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.906810 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.906826 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:21Z","lastTransitionTime":"2025-12-01T21:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.907968 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.934873 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.957385 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.984824 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:16Z\\\",\\\"message\\\":\\\"/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:16.788564 6247 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789027 6247 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789410 6247 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 21:35:16.790116 6247 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:16.790151 6247 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:16.790158 6247 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:16.790175 6247 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 21:35:16.790254 6247 factory.go:656] Stopping watch factory\\\\nI1201 21:35:16.790277 6247 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:16.790304 6247 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 21:35:16.790317 6247 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:16.790326 6247 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:21 crc kubenswrapper[4857]: I1201 21:35:21.996254 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.008275 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.011396 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.011680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.011769 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.011884 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.012066 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:22Z","lastTransitionTime":"2025-12-01T21:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.022489 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.035441 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.045253 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.057263 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.073503 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.083823 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.102580 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.115426 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.115968 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.115982 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.116005 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.116017 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:22Z","lastTransitionTime":"2025-12-01T21:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.119687 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.219358 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.219433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.219452 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.219476 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.219493 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:22Z","lastTransitionTime":"2025-12-01T21:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.323544 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.323671 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.323701 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.323730 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.323750 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:22Z","lastTransitionTime":"2025-12-01T21:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.427557 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.427625 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.427637 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.427656 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.427669 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:22Z","lastTransitionTime":"2025-12-01T21:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.530563 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.530645 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.530664 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.530692 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.530720 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:22Z","lastTransitionTime":"2025-12-01T21:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.634133 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.634194 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.634205 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.634225 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.634246 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:22Z","lastTransitionTime":"2025-12-01T21:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.737020 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.737142 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.737161 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.737188 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.737208 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:22Z","lastTransitionTime":"2025-12-01T21:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.834690 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:22 crc kubenswrapper[4857]: E1201 21:35:22.835103 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.834764 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:22 crc kubenswrapper[4857]: E1201 21:35:22.835396 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.840687 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.840740 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.840757 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.840781 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.840800 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:22Z","lastTransitionTime":"2025-12-01T21:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.944107 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.944489 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.944715 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.944951 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:22 crc kubenswrapper[4857]: I1201 21:35:22.945255 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:22Z","lastTransitionTime":"2025-12-01T21:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.048095 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.048125 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.048133 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.048147 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.048156 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:23Z","lastTransitionTime":"2025-12-01T21:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.151156 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.151204 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.151220 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.151244 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.151262 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:23Z","lastTransitionTime":"2025-12-01T21:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.254286 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.254344 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.254362 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.254388 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.254406 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:23Z","lastTransitionTime":"2025-12-01T21:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.357806 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.357869 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.357887 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.357913 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.357932 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:23Z","lastTransitionTime":"2025-12-01T21:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.460512 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.460597 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.460626 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.460655 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.460679 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:23Z","lastTransitionTime":"2025-12-01T21:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.563923 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.563969 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.563983 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.564002 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.564035 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:23Z","lastTransitionTime":"2025-12-01T21:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.667500 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.667564 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.667581 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.667611 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.667632 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:23Z","lastTransitionTime":"2025-12-01T21:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.680466 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:23 crc kubenswrapper[4857]: E1201 21:35:23.680643 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:23 crc kubenswrapper[4857]: E1201 21:35:23.680733 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs podName:485d0d95-8002-4bc3-8935-8b8f4ab5b2b3 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:31.680708193 +0000 UTC m=+50.170770550 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs") pod "network-metrics-daemon-n2fsz" (UID: "485d0d95-8002-4bc3-8935-8b8f4ab5b2b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.770651 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.770696 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.770709 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.770726 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.770738 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:23Z","lastTransitionTime":"2025-12-01T21:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.833982 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.833985 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:23 crc kubenswrapper[4857]: E1201 21:35:23.834215 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:23 crc kubenswrapper[4857]: E1201 21:35:23.834409 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.874129 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.874220 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.874246 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.874282 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.874304 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:23Z","lastTransitionTime":"2025-12-01T21:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.978101 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.978163 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.978179 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.978205 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:23 crc kubenswrapper[4857]: I1201 21:35:23.978227 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:23Z","lastTransitionTime":"2025-12-01T21:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.081410 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.081481 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.081498 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.081527 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.081552 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:24Z","lastTransitionTime":"2025-12-01T21:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.184600 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.184652 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.184671 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.184693 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.184710 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:24Z","lastTransitionTime":"2025-12-01T21:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.287096 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.287157 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.287177 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.287203 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.287223 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:24Z","lastTransitionTime":"2025-12-01T21:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.389945 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.389982 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.389991 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.390008 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.390018 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:24Z","lastTransitionTime":"2025-12-01T21:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.493690 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.493757 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.493779 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.493812 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.493840 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:24Z","lastTransitionTime":"2025-12-01T21:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.596844 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.596915 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.596939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.596969 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.596992 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:24Z","lastTransitionTime":"2025-12-01T21:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.700122 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.700185 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.700208 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.700238 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.700262 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:24Z","lastTransitionTime":"2025-12-01T21:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.802869 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.802912 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.802924 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.802940 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.802951 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:24Z","lastTransitionTime":"2025-12-01T21:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.834425 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.834513 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:24 crc kubenswrapper[4857]: E1201 21:35:24.834659 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:24 crc kubenswrapper[4857]: E1201 21:35:24.834732 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.906027 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.906200 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.906222 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.906250 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:24 crc kubenswrapper[4857]: I1201 21:35:24.906269 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:24Z","lastTransitionTime":"2025-12-01T21:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.009536 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.009582 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.009600 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.009625 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.009642 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.113100 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.113154 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.113171 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.113196 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.113214 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.198431 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.198491 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.198509 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.198533 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.198552 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: E1201 21:35:25.218666 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.223705 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.223750 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.223767 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.223790 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.223805 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: E1201 21:35:25.241944 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.247368 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.247417 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.247435 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.247456 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.247473 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: E1201 21:35:25.267278 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.272841 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.272876 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.272885 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.272919 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.272931 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: E1201 21:35:25.291216 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.301319 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.301385 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.301403 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.301433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.301453 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: E1201 21:35:25.320276 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:25 crc kubenswrapper[4857]: E1201 21:35:25.320581 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.322900 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.322993 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.323029 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.323101 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.323129 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.426575 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.426642 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.426661 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.426693 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.426713 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.530196 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.530251 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.530269 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.530297 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.530316 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.633727 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.633788 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.633806 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.633830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.633850 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.736975 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.737029 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.737073 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.737096 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.737112 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.834946 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.835072 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:25 crc kubenswrapper[4857]: E1201 21:35:25.835218 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:25 crc kubenswrapper[4857]: E1201 21:35:25.835349 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.841686 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.841749 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.841770 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.841798 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.841816 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.946248 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.946317 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.946336 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.946360 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:25 crc kubenswrapper[4857]: I1201 21:35:25.946377 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:25Z","lastTransitionTime":"2025-12-01T21:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.048955 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.049073 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.049102 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.049134 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.049156 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:26Z","lastTransitionTime":"2025-12-01T21:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.152287 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.152355 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.152373 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.152412 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.152431 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:26Z","lastTransitionTime":"2025-12-01T21:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.256068 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.256130 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.256147 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.256172 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.256191 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:26Z","lastTransitionTime":"2025-12-01T21:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.359326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.359392 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.359411 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.359436 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.359454 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:26Z","lastTransitionTime":"2025-12-01T21:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.462317 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.462371 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.462379 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.462395 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.462405 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:26Z","lastTransitionTime":"2025-12-01T21:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.565294 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.565367 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.565385 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.565417 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.565434 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:26Z","lastTransitionTime":"2025-12-01T21:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.667954 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.668027 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.668081 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.668106 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.668123 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:26Z","lastTransitionTime":"2025-12-01T21:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.770638 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.770747 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.770770 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.770795 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.770812 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:26Z","lastTransitionTime":"2025-12-01T21:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.834708 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.834723 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:26 crc kubenswrapper[4857]: E1201 21:35:26.834882 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:26 crc kubenswrapper[4857]: E1201 21:35:26.835017 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.873488 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.873603 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.873628 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.873657 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.873682 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:26Z","lastTransitionTime":"2025-12-01T21:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.977114 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.977184 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.977202 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.977236 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:26 crc kubenswrapper[4857]: I1201 21:35:26.977259 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:26Z","lastTransitionTime":"2025-12-01T21:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.080593 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.080660 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.080680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.080708 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.080726 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:27Z","lastTransitionTime":"2025-12-01T21:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.183917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.183978 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.183999 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.184023 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.184080 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:27Z","lastTransitionTime":"2025-12-01T21:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.287279 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.287351 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.287375 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.287407 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.287430 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:27Z","lastTransitionTime":"2025-12-01T21:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.391025 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.391148 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.391166 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.391192 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.391210 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:27Z","lastTransitionTime":"2025-12-01T21:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.494632 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.494696 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.494715 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.494741 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.494759 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:27Z","lastTransitionTime":"2025-12-01T21:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.597693 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.597763 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.597782 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.597807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.597825 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:27Z","lastTransitionTime":"2025-12-01T21:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.701208 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.701331 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.701399 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.701433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.701493 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:27Z","lastTransitionTime":"2025-12-01T21:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.804571 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.804638 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.804656 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.804684 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.804704 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:27Z","lastTransitionTime":"2025-12-01T21:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.834772 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.834818 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:27 crc kubenswrapper[4857]: E1201 21:35:27.834923 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:27 crc kubenswrapper[4857]: E1201 21:35:27.835081 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.907904 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.907972 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.907997 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.908025 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:27 crc kubenswrapper[4857]: I1201 21:35:27.908086 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:27Z","lastTransitionTime":"2025-12-01T21:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.010868 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.010929 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.010946 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.010972 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.010988 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:28Z","lastTransitionTime":"2025-12-01T21:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.113997 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.114066 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.114078 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.114094 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.114106 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:28Z","lastTransitionTime":"2025-12-01T21:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.216062 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.216101 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.216111 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.216124 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.216133 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:28Z","lastTransitionTime":"2025-12-01T21:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.319489 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.319538 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.319553 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.319605 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.319618 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:28Z","lastTransitionTime":"2025-12-01T21:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.422111 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.422506 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.422602 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.422712 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.422800 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:28Z","lastTransitionTime":"2025-12-01T21:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.526295 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.526344 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.526355 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.526373 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.526385 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:28Z","lastTransitionTime":"2025-12-01T21:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.628688 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.628723 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.628732 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.628744 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.628754 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:28Z","lastTransitionTime":"2025-12-01T21:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.731670 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.731760 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.731786 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.731820 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.731843 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:28Z","lastTransitionTime":"2025-12-01T21:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.833949 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.834012 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.834025 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.834079 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.834100 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.834118 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:28 crc kubenswrapper[4857]: E1201 21:35:28.834237 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.834095 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:28Z","lastTransitionTime":"2025-12-01T21:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:28 crc kubenswrapper[4857]: E1201 21:35:28.834334 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.938139 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.938215 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.938234 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.938262 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:28 crc kubenswrapper[4857]: I1201 21:35:28.938282 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:28Z","lastTransitionTime":"2025-12-01T21:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.041824 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.041876 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.041889 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.041911 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.041925 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:29Z","lastTransitionTime":"2025-12-01T21:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.145439 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.145512 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.145531 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.145559 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.145579 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:29Z","lastTransitionTime":"2025-12-01T21:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.249008 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.249104 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.249125 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.249157 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.249177 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:29Z","lastTransitionTime":"2025-12-01T21:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.352861 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.352927 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.352943 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.352963 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.352980 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:29Z","lastTransitionTime":"2025-12-01T21:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.404473 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.418276 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.423670 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.442524 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.456444 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.456497 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.456520 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.456549 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.456569 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:29Z","lastTransitionTime":"2025-12-01T21:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.460314 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.475739 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.492661 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.513228 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.540538 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:16Z\\\",\\\"message\\\":\\\"/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:16.788564 6247 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789027 6247 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789410 6247 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 21:35:16.790116 6247 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:16.790151 6247 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:16.790158 6247 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:16.790175 6247 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 21:35:16.790254 6247 factory.go:656] Stopping watch factory\\\\nI1201 21:35:16.790277 6247 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:16.790304 6247 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 21:35:16.790317 6247 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:16.790326 6247 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.556700 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.559574 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.559610 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.559623 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.559643 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.559658 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:29Z","lastTransitionTime":"2025-12-01T21:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.571711 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.596211 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.614862 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.629270 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.654814 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.662844 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.662907 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.662929 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.662957 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.662978 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:29Z","lastTransitionTime":"2025-12-01T21:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.672815 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.691432 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.713727 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.766266 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.766318 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.766329 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.766347 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.766359 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:29Z","lastTransitionTime":"2025-12-01T21:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.833950 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.834073 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:29 crc kubenswrapper[4857]: E1201 21:35:29.834122 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:29 crc kubenswrapper[4857]: E1201 21:35:29.834260 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.869294 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.869362 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.869379 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.869405 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.869426 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:29Z","lastTransitionTime":"2025-12-01T21:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.971991 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.972088 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.972100 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.972117 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:29 crc kubenswrapper[4857]: I1201 21:35:29.972129 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:29Z","lastTransitionTime":"2025-12-01T21:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.075397 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.075447 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.075460 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.075481 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.075494 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:30Z","lastTransitionTime":"2025-12-01T21:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.179419 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.179463 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.179473 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.179496 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.179511 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:30Z","lastTransitionTime":"2025-12-01T21:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.282359 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.282401 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.282412 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.282430 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.282442 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:30Z","lastTransitionTime":"2025-12-01T21:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.384975 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.385032 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.385072 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.385096 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.385114 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:30Z","lastTransitionTime":"2025-12-01T21:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.488255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.488322 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.488341 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.488368 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.488386 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:30Z","lastTransitionTime":"2025-12-01T21:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.592265 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.592329 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.592353 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.592379 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.592397 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:30Z","lastTransitionTime":"2025-12-01T21:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.695535 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.695619 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.695649 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.695681 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.695705 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:30Z","lastTransitionTime":"2025-12-01T21:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.798448 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.798509 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.798526 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.798550 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.798567 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:30Z","lastTransitionTime":"2025-12-01T21:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.833853 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.833886 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:30 crc kubenswrapper[4857]: E1201 21:35:30.834087 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:30 crc kubenswrapper[4857]: E1201 21:35:30.834194 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.901970 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.902036 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.902077 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.902102 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:30 crc kubenswrapper[4857]: I1201 21:35:30.902120 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:30Z","lastTransitionTime":"2025-12-01T21:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.005678 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.005776 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.005804 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.005833 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.005853 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:31Z","lastTransitionTime":"2025-12-01T21:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.109424 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.109493 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.109511 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.109543 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.109565 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:31Z","lastTransitionTime":"2025-12-01T21:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.212369 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.212438 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.212464 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.212498 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.212523 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:31Z","lastTransitionTime":"2025-12-01T21:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.315930 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.316003 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.316023 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.316086 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.316117 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:31Z","lastTransitionTime":"2025-12-01T21:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.418847 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.418896 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.418916 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.418937 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.418955 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:31Z","lastTransitionTime":"2025-12-01T21:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.522621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.522723 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.522744 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.522775 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.522799 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:31Z","lastTransitionTime":"2025-12-01T21:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.625446 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.625493 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.625503 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.625521 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.625540 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:31Z","lastTransitionTime":"2025-12-01T21:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.729165 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.729241 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.729259 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.729288 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.729307 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:31Z","lastTransitionTime":"2025-12-01T21:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.776312 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:31 crc kubenswrapper[4857]: E1201 21:35:31.776611 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:31 crc kubenswrapper[4857]: E1201 21:35:31.776732 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs podName:485d0d95-8002-4bc3-8935-8b8f4ab5b2b3 nodeName:}" failed. No retries permitted until 2025-12-01 21:35:47.776698232 +0000 UTC m=+66.266760589 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs") pod "network-metrics-daemon-n2fsz" (UID: "485d0d95-8002-4bc3-8935-8b8f4ab5b2b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.833126 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.833188 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.833208 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.833233 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.833252 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:31Z","lastTransitionTime":"2025-12-01T21:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.834147 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:31 crc kubenswrapper[4857]: E1201 21:35:31.834325 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.842357 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:31 crc kubenswrapper[4857]: E1201 21:35:31.842601 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.843214 4857 scope.go:117] "RemoveContainer" containerID="d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.873165 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.901153 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.920770 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e231d4d6-c755-4e41-bfe2-d65bd9765357\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.936820 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.936906 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.936933 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.936969 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.936997 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:31Z","lastTransitionTime":"2025-12-01T21:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.944883 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.962649 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.979072 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:31 crc kubenswrapper[4857]: I1201 21:35:31.990182 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.002647 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.014632 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.025325 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.034185 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.040016 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.040094 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.040110 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.040133 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.040149 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:32Z","lastTransitionTime":"2025-12-01T21:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.047328 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.064998 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:16Z\\\",\\\"message\\\":\\\"/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:16.788564 6247 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789027 6247 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789410 6247 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 21:35:16.790116 6247 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:16.790151 6247 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:16.790158 6247 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:16.790175 6247 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 21:35:16.790254 6247 factory.go:656] Stopping watch factory\\\\nI1201 21:35:16.790277 6247 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:16.790304 6247 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 21:35:16.790317 6247 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:16.790326 6247 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.077285 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.079887 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.080107 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:36:04.080084098 +0000 UTC m=+82.570146415 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.090265 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.100161 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.113371 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.142295 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.142722 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.142873 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.142966 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.143102 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:32Z","lastTransitionTime":"2025-12-01T21:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.181104 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.181191 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.181255 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.181329 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.181419 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.181438 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.181476 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.181489 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.181504 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.181513 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.181515 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.181590 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 21:36:04.181560175 +0000 UTC m=+82.671622522 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.181629 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 21:36:04.181612096 +0000 UTC m=+82.671674453 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.181656 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:36:04.181641107 +0000 UTC m=+82.671703464 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.182608 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.182714 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:36:04.182701021 +0000 UTC m=+82.672763338 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.246584 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.246654 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.246680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.246712 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.246734 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:32Z","lastTransitionTime":"2025-12-01T21:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.350089 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.350122 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.350132 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.350149 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.350159 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:32Z","lastTransitionTime":"2025-12-01T21:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.454986 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.455032 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.455065 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.455085 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.455099 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:32Z","lastTransitionTime":"2025-12-01T21:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.557996 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.558035 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.558068 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.558089 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.558102 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:32Z","lastTransitionTime":"2025-12-01T21:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.661068 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.661109 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.661120 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.661139 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.661150 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:32Z","lastTransitionTime":"2025-12-01T21:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.799321 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.799375 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.799390 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.799420 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.799436 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:32Z","lastTransitionTime":"2025-12-01T21:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.834658 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.834782 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.834869 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:32 crc kubenswrapper[4857]: E1201 21:35:32.835112 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.901975 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.902034 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.902093 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.902170 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:32 crc kubenswrapper[4857]: I1201 21:35:32.902189 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:32Z","lastTransitionTime":"2025-12-01T21:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.024357 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.024420 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.024431 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.024452 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.024466 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:33Z","lastTransitionTime":"2025-12-01T21:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.127680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.127745 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.127763 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.127787 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.127804 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:33Z","lastTransitionTime":"2025-12-01T21:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.229913 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.229973 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.229991 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.230015 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.230032 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:33Z","lastTransitionTime":"2025-12-01T21:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.230636 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/1.log" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.234606 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692"} Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.236387 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.258815 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.282998 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.301505 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.323853 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.333079 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.333235 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.333329 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.333444 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.333540 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:33Z","lastTransitionTime":"2025-12-01T21:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.345640 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.370001 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:16Z\\\",\\\"message\\\":\\\"/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:16.788564 6247 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789027 6247 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789410 6247 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 21:35:16.790116 6247 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:16.790151 6247 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:16.790158 6247 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:16.790175 6247 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 21:35:16.790254 6247 factory.go:656] Stopping watch factory\\\\nI1201 21:35:16.790277 6247 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:16.790304 6247 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 21:35:16.790317 6247 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:16.790326 6247 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.383666 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.401249 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.426808 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.436218 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.436287 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.436301 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.436321 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.436334 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:33Z","lastTransitionTime":"2025-12-01T21:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.449291 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.471797 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.494027 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.518605 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.539936 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.539997 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.540022 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.540101 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.540127 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:33Z","lastTransitionTime":"2025-12-01T21:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.541650 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.563926 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.592720 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.609620 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e231d4d6-c755-4e41-bfe2-d65bd9765357\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.642733 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.643218 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.643358 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.643539 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.643703 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:33Z","lastTransitionTime":"2025-12-01T21:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.747220 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.747280 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.747300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.747326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.747342 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:33Z","lastTransitionTime":"2025-12-01T21:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.834481 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:33 crc kubenswrapper[4857]: E1201 21:35:33.835293 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.834616 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:33 crc kubenswrapper[4857]: E1201 21:35:33.835718 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.850201 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.850249 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.850269 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.850290 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.850309 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:33Z","lastTransitionTime":"2025-12-01T21:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.954924 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.954978 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.954997 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.955021 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:33 crc kubenswrapper[4857]: I1201 21:35:33.955070 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:33Z","lastTransitionTime":"2025-12-01T21:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.058108 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.058171 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.058190 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.058219 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.058237 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:34Z","lastTransitionTime":"2025-12-01T21:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.162682 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.163189 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.163418 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.163613 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.163868 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:34Z","lastTransitionTime":"2025-12-01T21:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.244551 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/2.log" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.246242 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/1.log" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.251648 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692" exitCode=1 Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.251850 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692"} Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.252018 4857 scope.go:117] "RemoveContainer" containerID="d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.253585 4857 scope.go:117] "RemoveContainer" containerID="7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692" Dec 01 21:35:34 crc kubenswrapper[4857]: E1201 21:35:34.253888 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.267508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.267812 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.268088 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.268338 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.268598 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:34Z","lastTransitionTime":"2025-12-01T21:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.283632 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.304158 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.325194 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.348815 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.367849 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e231d4d6-c755-4e41-bfe2-d65bd9765357\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.377422 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.377631 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.377715 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.377805 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.377952 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:34Z","lastTransitionTime":"2025-12-01T21:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.389925 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.411690 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.430172 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.445690 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.464520 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.481273 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.481337 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.481348 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.481365 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.481376 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:34Z","lastTransitionTime":"2025-12-01T21:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.497628 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3b00ae08bd967f9135556ff9904043295a71235cada5c7fe0559daad5606a0a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:16Z\\\",\\\"message\\\":\\\"/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:16.788564 6247 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789027 6247 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:16.789410 6247 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 21:35:16.790116 6247 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:16.790151 6247 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:16.790158 6247 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:16.790175 6247 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 21:35:16.790254 6247 factory.go:656] Stopping watch factory\\\\nI1201 21:35:16.790277 6247 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:16.790304 6247 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 21:35:16.790317 6247 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:16.790326 6247 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:33Z\\\",\\\"message\\\":\\\" 21:35:33.310121 6430 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:33.310216 6430 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310592 6430 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:33.310642 6430 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310687 6430 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.311649 6430 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:33.311689 6430 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:33.311774 6430 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:33.311807 6430 factory.go:656] Stopping watch factory\\\\nI1201 21:35:33.311829 6430 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:33.311881 6430 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:33.311899 6430 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.512251 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.531828 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.546892 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.564705 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.578712 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.592337 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.592403 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.592420 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.592442 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.592461 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:34Z","lastTransitionTime":"2025-12-01T21:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.596916 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:34Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.695311 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.695373 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.695391 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.695417 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.695435 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:34Z","lastTransitionTime":"2025-12-01T21:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.797976 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.798028 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.798088 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.798114 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.798135 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:34Z","lastTransitionTime":"2025-12-01T21:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.834773 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.834830 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:34 crc kubenswrapper[4857]: E1201 21:35:34.835087 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:34 crc kubenswrapper[4857]: E1201 21:35:34.835250 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.901592 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.901638 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.901649 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.901666 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:34 crc kubenswrapper[4857]: I1201 21:35:34.901677 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:34Z","lastTransitionTime":"2025-12-01T21:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.005958 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.006357 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.006500 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.006647 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.006908 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.110398 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.110471 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.110485 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.110513 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.110531 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.214832 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.214909 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.214928 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.214959 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.214984 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.259381 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/2.log" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.264645 4857 scope.go:117] "RemoveContainer" containerID="7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692" Dec 01 21:35:35 crc kubenswrapper[4857]: E1201 21:35:35.264926 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.290689 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.311669 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.317721 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.317776 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.317797 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.317833 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.317857 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.334598 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.356348 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.377034 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.397807 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.420767 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.420839 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.420859 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.420891 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.420913 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.431320 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:33Z\\\",\\\"message\\\":\\\" 21:35:33.310121 6430 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:33.310216 6430 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310592 6430 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:33.310642 6430 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310687 6430 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.311649 6430 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:33.311689 6430 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:33.311774 6430 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:33.311807 6430 factory.go:656] Stopping watch factory\\\\nI1201 21:35:33.311829 6430 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:33.311881 6430 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:33.311899 6430 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.444168 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.457126 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.476797 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.493458 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.506688 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.520134 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.520206 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.520227 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.520254 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.520274 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.523529 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e231d4d6-c755-4e41-bfe2-d65bd9765357\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: E1201 21:35:35.541419 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.546270 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.546773 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.546881 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.546962 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.547105 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.547204 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.564997 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: E1201 21:35:35.570088 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.575176 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.575332 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.575450 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.575558 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.575645 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.586847 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: E1201 21:35:35.592864 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.597799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.597848 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.597869 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.597896 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.597914 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.604939 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: E1201 21:35:35.613912 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.618651 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.618709 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.618728 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.618753 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.618771 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: E1201 21:35:35.635117 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:35Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:35 crc kubenswrapper[4857]: E1201 21:35:35.635356 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.637329 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.637391 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.637416 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.637448 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.637469 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.740955 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.741029 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.741092 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.741123 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.741145 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.834185 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:35 crc kubenswrapper[4857]: E1201 21:35:35.834339 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.834391 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:35 crc kubenswrapper[4857]: E1201 21:35:35.834451 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.857102 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.857162 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.857175 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.857189 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.857201 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.960537 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.960589 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.960603 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.960623 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:35 crc kubenswrapper[4857]: I1201 21:35:35.960639 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:35Z","lastTransitionTime":"2025-12-01T21:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.063681 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.063756 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.063769 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.063789 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.063802 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:36Z","lastTransitionTime":"2025-12-01T21:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.167602 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.167678 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.167699 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.167734 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.167755 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:36Z","lastTransitionTime":"2025-12-01T21:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.270717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.270789 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.270803 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.270821 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.270834 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:36Z","lastTransitionTime":"2025-12-01T21:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.374510 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.374568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.374581 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.374604 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.374619 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:36Z","lastTransitionTime":"2025-12-01T21:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.477630 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.477713 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.477736 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.477766 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.477788 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:36Z","lastTransitionTime":"2025-12-01T21:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.580266 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.580319 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.580339 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.580364 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.580382 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:36Z","lastTransitionTime":"2025-12-01T21:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.683145 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.683189 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.683204 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.683225 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.683239 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:36Z","lastTransitionTime":"2025-12-01T21:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.785734 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.785778 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.785791 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.785812 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.785823 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:36Z","lastTransitionTime":"2025-12-01T21:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.834874 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.834970 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:36 crc kubenswrapper[4857]: E1201 21:35:36.835125 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:36 crc kubenswrapper[4857]: E1201 21:35:36.835233 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.888011 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.888092 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.888111 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.888131 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.888144 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:36Z","lastTransitionTime":"2025-12-01T21:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.991585 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.991658 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.991682 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.991705 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:36 crc kubenswrapper[4857]: I1201 21:35:36.991723 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:36Z","lastTransitionTime":"2025-12-01T21:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.094744 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.094799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.094815 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.094837 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.094854 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:37Z","lastTransitionTime":"2025-12-01T21:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.197971 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.198086 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.198115 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.198149 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.198170 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:37Z","lastTransitionTime":"2025-12-01T21:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.300863 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.300924 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.300939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.300959 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.300974 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:37Z","lastTransitionTime":"2025-12-01T21:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.404327 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.404419 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.404472 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.404494 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.404515 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:37Z","lastTransitionTime":"2025-12-01T21:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.507466 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.507543 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.507567 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.507597 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.507620 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:37Z","lastTransitionTime":"2025-12-01T21:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.610038 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.610160 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.610185 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.610211 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.610229 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:37Z","lastTransitionTime":"2025-12-01T21:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.713485 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.713550 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.713570 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.713599 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.713631 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:37Z","lastTransitionTime":"2025-12-01T21:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.816622 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.816685 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.816713 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.816737 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.816755 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:37Z","lastTransitionTime":"2025-12-01T21:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.834179 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.834188 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:37 crc kubenswrapper[4857]: E1201 21:35:37.834342 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:37 crc kubenswrapper[4857]: E1201 21:35:37.834554 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.920030 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.920160 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.920186 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.920216 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:37 crc kubenswrapper[4857]: I1201 21:35:37.920248 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:37Z","lastTransitionTime":"2025-12-01T21:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.023203 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.023262 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.023279 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.023302 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.023319 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:38Z","lastTransitionTime":"2025-12-01T21:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.126932 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.127010 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.127073 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.127100 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.127118 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:38Z","lastTransitionTime":"2025-12-01T21:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.230656 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.230711 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.230724 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.230743 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.230754 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:38Z","lastTransitionTime":"2025-12-01T21:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.333830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.333931 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.333957 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.334027 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.334107 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:38Z","lastTransitionTime":"2025-12-01T21:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.437875 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.437943 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.437967 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.438003 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.438135 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:38Z","lastTransitionTime":"2025-12-01T21:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.541555 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.541612 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.541630 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.541653 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.541672 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:38Z","lastTransitionTime":"2025-12-01T21:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.644708 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.644780 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.644803 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.644835 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.644856 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:38Z","lastTransitionTime":"2025-12-01T21:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.747544 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.747593 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.747614 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.747639 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.747660 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:38Z","lastTransitionTime":"2025-12-01T21:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.834317 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.834355 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:38 crc kubenswrapper[4857]: E1201 21:35:38.834498 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:38 crc kubenswrapper[4857]: E1201 21:35:38.834692 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.852939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.853000 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.853018 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.853079 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.853099 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:38Z","lastTransitionTime":"2025-12-01T21:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.956592 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.956636 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.956648 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.956666 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:38 crc kubenswrapper[4857]: I1201 21:35:38.956679 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:38Z","lastTransitionTime":"2025-12-01T21:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.059990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.060089 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.060108 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.060135 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.060155 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:39Z","lastTransitionTime":"2025-12-01T21:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.163558 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.163621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.163639 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.163663 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.163683 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:39Z","lastTransitionTime":"2025-12-01T21:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.266627 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.266682 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.266699 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.266721 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.266737 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:39Z","lastTransitionTime":"2025-12-01T21:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.370112 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.370422 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.370442 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.370468 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.370487 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:39Z","lastTransitionTime":"2025-12-01T21:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.473267 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.473335 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.473353 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.473381 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.473399 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:39Z","lastTransitionTime":"2025-12-01T21:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.576616 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.576711 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.576731 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.576761 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.576804 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:39Z","lastTransitionTime":"2025-12-01T21:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.679571 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.679827 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.679911 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.679997 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.680133 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:39Z","lastTransitionTime":"2025-12-01T21:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.783321 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.783391 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.783410 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.783437 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.783460 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:39Z","lastTransitionTime":"2025-12-01T21:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.836940 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:39 crc kubenswrapper[4857]: E1201 21:35:39.837160 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.837813 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:39 crc kubenswrapper[4857]: E1201 21:35:39.837927 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.886054 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.886104 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.886116 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.886132 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.886146 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:39Z","lastTransitionTime":"2025-12-01T21:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.988912 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.988964 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.988980 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.989005 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:39 crc kubenswrapper[4857]: I1201 21:35:39.989024 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:39Z","lastTransitionTime":"2025-12-01T21:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.091939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.091998 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.092017 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.092072 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.092091 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:40Z","lastTransitionTime":"2025-12-01T21:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.196185 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.196251 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.196274 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.196305 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.196328 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:40Z","lastTransitionTime":"2025-12-01T21:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.299223 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.299276 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.299293 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.299322 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.299342 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:40Z","lastTransitionTime":"2025-12-01T21:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.402105 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.402168 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.402184 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.402213 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.402250 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:40Z","lastTransitionTime":"2025-12-01T21:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.505291 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.505365 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.505393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.505420 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.505441 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:40Z","lastTransitionTime":"2025-12-01T21:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.608802 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.608848 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.608858 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.608874 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.608887 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:40Z","lastTransitionTime":"2025-12-01T21:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.711936 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.711982 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.711992 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.712007 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.712016 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:40Z","lastTransitionTime":"2025-12-01T21:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.815013 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.815107 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.815126 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.815150 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.815168 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:40Z","lastTransitionTime":"2025-12-01T21:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.834441 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:40 crc kubenswrapper[4857]: E1201 21:35:40.834599 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.834437 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:40 crc kubenswrapper[4857]: E1201 21:35:40.834725 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.918602 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.918647 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.918656 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.918672 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:40 crc kubenswrapper[4857]: I1201 21:35:40.918685 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:40Z","lastTransitionTime":"2025-12-01T21:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.021704 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.021761 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.021778 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.021806 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.021829 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:41Z","lastTransitionTime":"2025-12-01T21:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.125596 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.125662 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.125678 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.125703 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.125720 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:41Z","lastTransitionTime":"2025-12-01T21:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.228945 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.229005 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.229028 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.229090 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.229110 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:41Z","lastTransitionTime":"2025-12-01T21:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.331645 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.331700 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.331717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.331743 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.331760 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:41Z","lastTransitionTime":"2025-12-01T21:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.435230 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.435284 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.435301 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.435327 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.435347 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:41Z","lastTransitionTime":"2025-12-01T21:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.537914 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.537971 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.537988 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.538012 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.538031 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:41Z","lastTransitionTime":"2025-12-01T21:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.641388 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.641447 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.641468 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.641501 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.641524 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:41Z","lastTransitionTime":"2025-12-01T21:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.744856 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.744914 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.744932 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.744958 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.744977 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:41Z","lastTransitionTime":"2025-12-01T21:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.834160 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:41 crc kubenswrapper[4857]: E1201 21:35:41.834380 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.834437 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:41 crc kubenswrapper[4857]: E1201 21:35:41.836183 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.849066 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.849103 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.849117 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.849138 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.849153 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:41Z","lastTransitionTime":"2025-12-01T21:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.859315 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:41Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.879744 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:41Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.897321 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:41Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.943666 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:41Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.954344 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.954382 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.954393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.954409 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.954421 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:41Z","lastTransitionTime":"2025-12-01T21:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.969552 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:41Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:41 crc kubenswrapper[4857]: I1201 21:35:41.981768 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:41Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.003869 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:33Z\\\",\\\"message\\\":\\\" 21:35:33.310121 6430 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:33.310216 6430 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310592 6430 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:33.310642 6430 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310687 6430 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.311649 6430 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:33.311689 6430 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:33.311774 6430 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:33.311807 6430 factory.go:656] Stopping watch factory\\\\nI1201 21:35:33.311829 6430 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:33.311881 6430 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:33.311899 6430 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.018567 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.033766 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.057306 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.057672 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.057855 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.058107 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.057636 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.058327 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:42Z","lastTransitionTime":"2025-12-01T21:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.079269 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.093982 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.107834 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e231d4d6-c755-4e41-bfe2-d65bd9765357\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.128148 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.146108 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.161530 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.161600 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.161628 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.161659 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.161699 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:42Z","lastTransitionTime":"2025-12-01T21:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.167677 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.182807 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.264730 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.264791 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.264810 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.264839 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.264862 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:42Z","lastTransitionTime":"2025-12-01T21:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.368094 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.368481 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.368640 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.368804 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.368956 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:42Z","lastTransitionTime":"2025-12-01T21:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.474266 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.474335 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.474354 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.474426 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.475117 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:42Z","lastTransitionTime":"2025-12-01T21:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.578555 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.578641 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.578669 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.578701 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.578724 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:42Z","lastTransitionTime":"2025-12-01T21:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.682218 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.682277 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.682295 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.682321 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.682340 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:42Z","lastTransitionTime":"2025-12-01T21:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.786003 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.786092 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.786114 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.786139 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.786157 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:42Z","lastTransitionTime":"2025-12-01T21:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.834834 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.834900 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:42 crc kubenswrapper[4857]: E1201 21:35:42.835021 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:42 crc kubenswrapper[4857]: E1201 21:35:42.835174 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.889249 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.889316 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.889333 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.889358 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.889378 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:42Z","lastTransitionTime":"2025-12-01T21:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.992264 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.992328 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.992347 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.992374 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:42 crc kubenswrapper[4857]: I1201 21:35:42.992393 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:42Z","lastTransitionTime":"2025-12-01T21:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.095188 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.095266 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.095292 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.095323 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.095348 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:43Z","lastTransitionTime":"2025-12-01T21:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.199100 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.199172 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.199190 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.199216 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.199234 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:43Z","lastTransitionTime":"2025-12-01T21:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.301996 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.302062 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.302079 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.302098 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.302114 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:43Z","lastTransitionTime":"2025-12-01T21:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.404810 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.404906 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.404928 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.404959 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.404982 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:43Z","lastTransitionTime":"2025-12-01T21:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.507611 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.507686 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.507710 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.507739 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.507760 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:43Z","lastTransitionTime":"2025-12-01T21:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.610634 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.610711 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.610725 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.610741 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.610753 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:43Z","lastTransitionTime":"2025-12-01T21:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.713830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.713880 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.713891 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.713907 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.713917 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:43Z","lastTransitionTime":"2025-12-01T21:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.816858 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.816900 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.816913 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.816929 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.816941 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:43Z","lastTransitionTime":"2025-12-01T21:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.834517 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:43 crc kubenswrapper[4857]: E1201 21:35:43.834652 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.834673 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:43 crc kubenswrapper[4857]: E1201 21:35:43.834750 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.920187 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.920260 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.920283 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.920312 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:43 crc kubenswrapper[4857]: I1201 21:35:43.920331 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:43Z","lastTransitionTime":"2025-12-01T21:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.024135 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.024206 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.024233 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.024266 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.024289 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:44Z","lastTransitionTime":"2025-12-01T21:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.128350 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.128448 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.128465 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.128488 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.128505 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:44Z","lastTransitionTime":"2025-12-01T21:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.231584 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.231640 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.231659 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.231683 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.231701 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:44Z","lastTransitionTime":"2025-12-01T21:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.334232 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.334287 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.334300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.334323 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.334340 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:44Z","lastTransitionTime":"2025-12-01T21:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.437174 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.437300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.437321 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.437351 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.437371 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:44Z","lastTransitionTime":"2025-12-01T21:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.540285 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.540395 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.540414 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.540438 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.540455 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:44Z","lastTransitionTime":"2025-12-01T21:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.642950 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.643007 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.643026 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.643084 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.643123 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:44Z","lastTransitionTime":"2025-12-01T21:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.745875 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.745946 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.745971 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.745999 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.746019 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:44Z","lastTransitionTime":"2025-12-01T21:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.834073 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.834092 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:44 crc kubenswrapper[4857]: E1201 21:35:44.834387 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:44 crc kubenswrapper[4857]: E1201 21:35:44.834741 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.849439 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.849483 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.849512 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.849535 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.849548 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:44Z","lastTransitionTime":"2025-12-01T21:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.952313 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.952395 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.952420 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.952450 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:44 crc kubenswrapper[4857]: I1201 21:35:44.952474 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:44Z","lastTransitionTime":"2025-12-01T21:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.054713 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.055158 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.055177 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.055200 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.055221 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.157668 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.157741 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.157766 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.157801 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.157825 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.260655 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.260702 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.260715 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.260731 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.260741 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.365447 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.365541 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.365556 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.365573 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.365586 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.468153 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.468199 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.468210 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.468224 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.468234 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.570955 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.570989 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.570998 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.571011 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.571020 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.673752 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.673811 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.673828 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.673850 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.673867 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.780685 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.780746 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.780764 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.780789 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.780806 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.832601 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.832717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.833032 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.833318 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.833340 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.834839 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.834867 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:45 crc kubenswrapper[4857]: E1201 21:35:45.835018 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:45 crc kubenswrapper[4857]: E1201 21:35:45.835164 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:45 crc kubenswrapper[4857]: E1201 21:35:45.852692 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.856551 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.856606 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.856631 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.856654 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.856669 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: E1201 21:35:45.871006 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.874970 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.875055 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.875074 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.875096 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.875113 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: E1201 21:35:45.895694 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.899546 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.899621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.899646 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.899673 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.899690 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: E1201 21:35:45.912098 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.916216 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.916250 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.916262 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.916280 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.916293 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:45 crc kubenswrapper[4857]: E1201 21:35:45.932747 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:45 crc kubenswrapper[4857]: E1201 21:35:45.932863 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.934621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.934649 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.934659 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.934676 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:45 crc kubenswrapper[4857]: I1201 21:35:45.934691 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:45Z","lastTransitionTime":"2025-12-01T21:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.037739 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.037797 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.037810 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.037834 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.037847 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:46Z","lastTransitionTime":"2025-12-01T21:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.140184 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.140228 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.140247 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.140270 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.140287 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:46Z","lastTransitionTime":"2025-12-01T21:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.243432 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.243477 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.243489 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.243506 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.243519 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:46Z","lastTransitionTime":"2025-12-01T21:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.346606 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.346647 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.346655 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.346671 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.346681 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:46Z","lastTransitionTime":"2025-12-01T21:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.449222 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.449267 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.449279 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.449297 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.449311 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:46Z","lastTransitionTime":"2025-12-01T21:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.551803 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.551853 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.551867 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.551886 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.551898 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:46Z","lastTransitionTime":"2025-12-01T21:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.654006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.654089 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.654102 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.654122 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.654135 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:46Z","lastTransitionTime":"2025-12-01T21:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.756562 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.756605 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.756616 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.756631 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.756640 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:46Z","lastTransitionTime":"2025-12-01T21:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.833966 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.834008 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:46 crc kubenswrapper[4857]: E1201 21:35:46.834213 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:46 crc kubenswrapper[4857]: E1201 21:35:46.834423 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.859118 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.859144 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.859152 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.859165 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.859175 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:46Z","lastTransitionTime":"2025-12-01T21:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.961231 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.961259 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.961267 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.961279 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:46 crc kubenswrapper[4857]: I1201 21:35:46.961287 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:46Z","lastTransitionTime":"2025-12-01T21:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.064227 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.064300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.064317 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.064339 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.064352 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:47Z","lastTransitionTime":"2025-12-01T21:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.166895 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.166931 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.166939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.166957 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.166968 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:47Z","lastTransitionTime":"2025-12-01T21:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.269433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.269500 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.269518 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.269546 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.269564 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:47Z","lastTransitionTime":"2025-12-01T21:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.371357 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.371398 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.371408 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.371422 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.371434 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:47Z","lastTransitionTime":"2025-12-01T21:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.473788 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.473883 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.473918 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.473950 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.473974 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:47Z","lastTransitionTime":"2025-12-01T21:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.577439 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.577530 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.577556 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.577588 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.577611 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:47Z","lastTransitionTime":"2025-12-01T21:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.681161 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.681219 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.681232 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.681256 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.681269 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:47Z","lastTransitionTime":"2025-12-01T21:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.783724 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.783763 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.783771 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.783785 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.783795 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:47Z","lastTransitionTime":"2025-12-01T21:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.834904 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:47 crc kubenswrapper[4857]: E1201 21:35:47.835056 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.835114 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:47 crc kubenswrapper[4857]: E1201 21:35:47.835852 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.836948 4857 scope.go:117] "RemoveContainer" containerID="7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692" Dec 01 21:35:47 crc kubenswrapper[4857]: E1201 21:35:47.837299 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.844556 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.866427 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:47 crc kubenswrapper[4857]: E1201 21:35:47.866521 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:47 crc kubenswrapper[4857]: E1201 21:35:47.866562 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs podName:485d0d95-8002-4bc3-8935-8b8f4ab5b2b3 nodeName:}" failed. No retries permitted until 2025-12-01 21:36:19.866548379 +0000 UTC m=+98.356610696 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs") pod "network-metrics-daemon-n2fsz" (UID: "485d0d95-8002-4bc3-8935-8b8f4ab5b2b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.885521 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.885560 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.885571 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.885585 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.885598 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:47Z","lastTransitionTime":"2025-12-01T21:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.988688 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.988724 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.988735 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.988755 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:47 crc kubenswrapper[4857]: I1201 21:35:47.988769 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:47Z","lastTransitionTime":"2025-12-01T21:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.091425 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.091488 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.091508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.091532 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.091549 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:48Z","lastTransitionTime":"2025-12-01T21:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.194342 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.194698 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.194881 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.195127 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.195313 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:48Z","lastTransitionTime":"2025-12-01T21:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.298122 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.298191 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.298205 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.298225 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.298239 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:48Z","lastTransitionTime":"2025-12-01T21:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.400951 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.401006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.401018 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.401058 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.401072 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:48Z","lastTransitionTime":"2025-12-01T21:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.503595 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.503651 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.503668 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.503693 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.503710 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:48Z","lastTransitionTime":"2025-12-01T21:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.606617 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.606683 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.606699 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.606721 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.606733 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:48Z","lastTransitionTime":"2025-12-01T21:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.708795 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.708886 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.708905 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.708929 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.708946 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:48Z","lastTransitionTime":"2025-12-01T21:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.812030 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.812159 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.812189 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.812224 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.812253 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:48Z","lastTransitionTime":"2025-12-01T21:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.834724 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.834843 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:48 crc kubenswrapper[4857]: E1201 21:35:48.834931 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:48 crc kubenswrapper[4857]: E1201 21:35:48.835092 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.915620 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.915672 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.915688 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.915709 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:48 crc kubenswrapper[4857]: I1201 21:35:48.915726 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:48Z","lastTransitionTime":"2025-12-01T21:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.018256 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.018293 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.018306 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.018330 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.018343 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:49Z","lastTransitionTime":"2025-12-01T21:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.121052 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.121096 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.121106 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.121120 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.121130 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:49Z","lastTransitionTime":"2025-12-01T21:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.223719 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.223768 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.223781 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.223800 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.223816 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:49Z","lastTransitionTime":"2025-12-01T21:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.325910 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.325969 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.325987 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.326010 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.326030 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:49Z","lastTransitionTime":"2025-12-01T21:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.428885 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.428946 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.428960 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.428995 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.429008 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:49Z","lastTransitionTime":"2025-12-01T21:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.531387 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.531422 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.531432 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.531448 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.531458 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:49Z","lastTransitionTime":"2025-12-01T21:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.633374 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.633416 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.633433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.633454 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.633472 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:49Z","lastTransitionTime":"2025-12-01T21:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.739724 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.739812 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.739833 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.739867 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.739888 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:49Z","lastTransitionTime":"2025-12-01T21:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.834874 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.834919 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:49 crc kubenswrapper[4857]: E1201 21:35:49.834999 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:49 crc kubenswrapper[4857]: E1201 21:35:49.835136 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.842352 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.842401 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.842414 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.842431 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.842445 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:49Z","lastTransitionTime":"2025-12-01T21:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.944659 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.944717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.944738 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.944767 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:49 crc kubenswrapper[4857]: I1201 21:35:49.944789 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:49Z","lastTransitionTime":"2025-12-01T21:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.047492 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.047544 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.047560 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.047580 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.047617 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:50Z","lastTransitionTime":"2025-12-01T21:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.150221 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.150258 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.150269 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.150286 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.150301 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:50Z","lastTransitionTime":"2025-12-01T21:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.252504 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.252545 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.252559 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.252575 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.252586 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:50Z","lastTransitionTime":"2025-12-01T21:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.355365 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.355541 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.355564 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.355634 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.355664 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:50Z","lastTransitionTime":"2025-12-01T21:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.458599 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.458655 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.458675 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.458699 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.458719 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:50Z","lastTransitionTime":"2025-12-01T21:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.562879 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.562917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.562931 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.562951 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.562964 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:50Z","lastTransitionTime":"2025-12-01T21:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.665049 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.665082 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.665092 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.665114 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.665126 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:50Z","lastTransitionTime":"2025-12-01T21:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.766855 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.766900 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.766913 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.766929 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.766941 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:50Z","lastTransitionTime":"2025-12-01T21:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.833936 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.834117 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:50 crc kubenswrapper[4857]: E1201 21:35:50.834121 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:50 crc kubenswrapper[4857]: E1201 21:35:50.834363 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.887402 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.887430 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.887439 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.887453 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.887462 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:50Z","lastTransitionTime":"2025-12-01T21:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.989626 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.989674 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.989684 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.989700 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:50 crc kubenswrapper[4857]: I1201 21:35:50.989712 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:50Z","lastTransitionTime":"2025-12-01T21:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.092207 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.092244 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.092255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.092269 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.092278 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:51Z","lastTransitionTime":"2025-12-01T21:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.194559 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.194639 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.194662 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.194692 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.194714 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:51Z","lastTransitionTime":"2025-12-01T21:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.297772 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.297841 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.297860 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.297893 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.297918 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:51Z","lastTransitionTime":"2025-12-01T21:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.401520 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.401591 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.401610 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.401637 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.401655 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:51Z","lastTransitionTime":"2025-12-01T21:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.505065 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.505103 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.505112 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.505131 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.505144 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:51Z","lastTransitionTime":"2025-12-01T21:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.608080 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.608117 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.608126 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.608141 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.608151 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:51Z","lastTransitionTime":"2025-12-01T21:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.711991 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.712070 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.712088 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.712114 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.712132 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:51Z","lastTransitionTime":"2025-12-01T21:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.814308 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.814356 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.814372 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.814390 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.814401 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:51Z","lastTransitionTime":"2025-12-01T21:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.834263 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.834325 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:51 crc kubenswrapper[4857]: E1201 21:35:51.834425 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:51 crc kubenswrapper[4857]: E1201 21:35:51.834511 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.848123 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.864107 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.888664 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:33Z\\\",\\\"message\\\":\\\" 21:35:33.310121 6430 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:33.310216 6430 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310592 6430 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:33.310642 6430 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310687 6430 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.311649 6430 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:33.311689 6430 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:33.311774 6430 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:33.311807 6430 factory.go:656] Stopping watch factory\\\\nI1201 21:35:33.311829 6430 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:33.311881 6430 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:33.311899 6430 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.900725 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.916005 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.916065 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.916075 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.916090 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.916100 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:51Z","lastTransitionTime":"2025-12-01T21:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.923596 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.970586 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:51 crc kubenswrapper[4857]: I1201 21:35:51.991019 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.004220 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116aa470-b443-48c1-91ce-bf63e232b25f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://005b5e00ddaba85767652a371cfdae9b98238fc5b1ec89beecdb2c18d9b2062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.019373 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.019429 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.019439 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.019455 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.019466 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:52Z","lastTransitionTime":"2025-12-01T21:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.023542 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.040912 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.056852 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e231d4d6-c755-4e41-bfe2-d65bd9765357\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.073397 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.086591 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.099425 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.111011 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.121744 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.121773 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.121802 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.121816 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.121825 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:52Z","lastTransitionTime":"2025-12-01T21:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.126627 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.149356 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.163825 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.224572 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.225081 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.225155 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.225238 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.225321 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:52Z","lastTransitionTime":"2025-12-01T21:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.327180 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.327231 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.327245 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.327264 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.327277 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:52Z","lastTransitionTime":"2025-12-01T21:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.552786 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.552820 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.552864 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.552877 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.552885 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:52Z","lastTransitionTime":"2025-12-01T21:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.655018 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.655087 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.655097 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.655112 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.655122 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:52Z","lastTransitionTime":"2025-12-01T21:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.758768 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.758813 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.758826 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.758836 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.758846 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:52Z","lastTransitionTime":"2025-12-01T21:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.834929 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.834929 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:52 crc kubenswrapper[4857]: E1201 21:35:52.835096 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:52 crc kubenswrapper[4857]: E1201 21:35:52.835225 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.861648 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.861695 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.861712 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.861733 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.861749 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:52Z","lastTransitionTime":"2025-12-01T21:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.964305 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.964364 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.964382 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.964409 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:52 crc kubenswrapper[4857]: I1201 21:35:52.964425 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:52Z","lastTransitionTime":"2025-12-01T21:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.066773 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.066822 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.066840 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.066893 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.066909 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:53Z","lastTransitionTime":"2025-12-01T21:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.169664 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.169744 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.169769 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.169799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.169825 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:53Z","lastTransitionTime":"2025-12-01T21:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.272260 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.272302 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.272314 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.272330 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.272341 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:53Z","lastTransitionTime":"2025-12-01T21:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.326191 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dfxtl_fc7d55e5-8085-40b7-8147-8508ee6c3f61/kube-multus/0.log" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.326254 4857 generic.go:334] "Generic (PLEG): container finished" podID="fc7d55e5-8085-40b7-8147-8508ee6c3f61" containerID="8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3" exitCode=1 Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.326303 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dfxtl" event={"ID":"fc7d55e5-8085-40b7-8147-8508ee6c3f61","Type":"ContainerDied","Data":"8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3"} Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.327567 4857 scope.go:117] "RemoveContainer" containerID="8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.356305 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.371957 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.376136 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.376185 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.376216 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.376239 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.376253 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:53Z","lastTransitionTime":"2025-12-01T21:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.396320 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.413606 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116aa470-b443-48c1-91ce-bf63e232b25f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://005b5e00ddaba85767652a371cfdae9b98238fc5b1ec89beecdb2c18d9b2062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.437744 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.455165 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:53Z\\\",\\\"message\\\":\\\"2025-12-01T21:35:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bba25fa0-e0dc-4aea-9563-b9ea666147ec\\\\n2025-12-01T21:35:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bba25fa0-e0dc-4aea-9563-b9ea666147ec to /host/opt/cni/bin/\\\\n2025-12-01T21:35:08Z [verbose] multus-daemon started\\\\n2025-12-01T21:35:08Z [verbose] Readiness Indicator file check\\\\n2025-12-01T21:35:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.469004 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e231d4d6-c755-4e41-bfe2-d65bd9765357\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.478353 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.478461 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.478539 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.478602 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.478661 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:53Z","lastTransitionTime":"2025-12-01T21:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.487586 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.506631 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.523116 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.539204 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.558077 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.576452 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.580978 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.581118 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.581180 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.581240 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.581316 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:53Z","lastTransitionTime":"2025-12-01T21:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.596659 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.612350 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.629192 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.655625 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:33Z\\\",\\\"message\\\":\\\" 21:35:33.310121 6430 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:33.310216 6430 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310592 6430 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:33.310642 6430 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310687 6430 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.311649 6430 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:33.311689 6430 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:33.311774 6430 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:33.311807 6430 factory.go:656] Stopping watch factory\\\\nI1201 21:35:33.311829 6430 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:33.311881 6430 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:33.311899 6430 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.670449 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:53Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.684295 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.684358 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.684377 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.684402 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.684421 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:53Z","lastTransitionTime":"2025-12-01T21:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.787908 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.787947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.787964 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.788008 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.788026 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:53Z","lastTransitionTime":"2025-12-01T21:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.834618 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.834693 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:53 crc kubenswrapper[4857]: E1201 21:35:53.834791 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:53 crc kubenswrapper[4857]: E1201 21:35:53.834994 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.890920 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.891171 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.891248 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.891310 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.891364 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:53Z","lastTransitionTime":"2025-12-01T21:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.995094 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.995138 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.995155 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.995178 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:53 crc kubenswrapper[4857]: I1201 21:35:53.995195 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:53Z","lastTransitionTime":"2025-12-01T21:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.097631 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.097670 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.097687 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.097709 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.097727 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:54Z","lastTransitionTime":"2025-12-01T21:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.200384 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.200427 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.200444 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.200467 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.200484 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:54Z","lastTransitionTime":"2025-12-01T21:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.303921 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.304114 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.304128 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.304141 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.304150 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:54Z","lastTransitionTime":"2025-12-01T21:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.333958 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dfxtl_fc7d55e5-8085-40b7-8147-8508ee6c3f61/kube-multus/0.log" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.334023 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dfxtl" event={"ID":"fc7d55e5-8085-40b7-8147-8508ee6c3f61","Type":"ContainerStarted","Data":"6e38c5755141156082b254a121520d58f44c5cef43ad63c3cadb2f1cf8e186e7"} Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.350581 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e231d4d6-c755-4e41-bfe2-d65bd9765357\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.366360 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.377262 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.389374 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.414585 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e38c5755141156082b254a121520d58f44c5cef43ad63c3cadb2f1cf8e186e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:53Z\\\",\\\"message\\\":\\\"2025-12-01T21:35:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bba25fa0-e0dc-4aea-9563-b9ea666147ec\\\\n2025-12-01T21:35:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bba25fa0-e0dc-4aea-9563-b9ea666147ec to /host/opt/cni/bin/\\\\n2025-12-01T21:35:08Z [verbose] multus-daemon started\\\\n2025-12-01T21:35:08Z [verbose] Readiness Indicator file check\\\\n2025-12-01T21:35:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.429273 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.445552 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.458114 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.470584 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.483057 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.483604 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.483649 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.483662 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.483680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.483692 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:54Z","lastTransitionTime":"2025-12-01T21:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.496888 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.514751 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:33Z\\\",\\\"message\\\":\\\" 21:35:33.310121 6430 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:33.310216 6430 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310592 6430 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:33.310642 6430 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310687 6430 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.311649 6430 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:33.311689 6430 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:33.311774 6430 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:33.311807 6430 factory.go:656] Stopping watch factory\\\\nI1201 21:35:33.311829 6430 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:33.311881 6430 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:33.311899 6430 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.525426 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.537678 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.550754 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.563607 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116aa470-b443-48c1-91ce-bf63e232b25f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://005b5e00ddaba85767652a371cfdae9b98238fc5b1ec89beecdb2c18d9b2062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.578389 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.588178 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.588212 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.588224 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.588239 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.588251 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:54Z","lastTransitionTime":"2025-12-01T21:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.590377 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:54Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.690632 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.690668 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.690680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.690697 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.690710 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:54Z","lastTransitionTime":"2025-12-01T21:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.793940 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.794031 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.794089 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.794136 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.794165 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:54Z","lastTransitionTime":"2025-12-01T21:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.834621 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.834621 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:54 crc kubenswrapper[4857]: E1201 21:35:54.835899 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:54 crc kubenswrapper[4857]: E1201 21:35:54.835899 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.897653 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.897715 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.897733 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.897757 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:54 crc kubenswrapper[4857]: I1201 21:35:54.897779 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:54Z","lastTransitionTime":"2025-12-01T21:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.000777 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.000832 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.000845 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.000866 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.000880 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:55Z","lastTransitionTime":"2025-12-01T21:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.103979 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.104036 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.104071 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.104105 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.104121 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:55Z","lastTransitionTime":"2025-12-01T21:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.206786 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.206855 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.206879 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.206907 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.206929 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:55Z","lastTransitionTime":"2025-12-01T21:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.309706 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.309757 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.309776 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.309800 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.309817 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:55Z","lastTransitionTime":"2025-12-01T21:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.412990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.413067 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.413087 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.413109 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.413125 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:55Z","lastTransitionTime":"2025-12-01T21:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.516247 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.516310 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.516334 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.516364 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.516385 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:55Z","lastTransitionTime":"2025-12-01T21:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.619455 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.619523 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.619541 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.619569 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.619586 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:55Z","lastTransitionTime":"2025-12-01T21:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.723175 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.723232 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.723249 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.723275 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.723292 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:55Z","lastTransitionTime":"2025-12-01T21:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.826412 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.826467 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.826485 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.826507 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.826523 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:55Z","lastTransitionTime":"2025-12-01T21:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.833874 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:55 crc kubenswrapper[4857]: E1201 21:35:55.834036 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.834312 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:55 crc kubenswrapper[4857]: E1201 21:35:55.834538 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.929603 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.929670 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.929690 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.929715 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:55 crc kubenswrapper[4857]: I1201 21:35:55.929740 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:55Z","lastTransitionTime":"2025-12-01T21:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.032729 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.032770 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.032778 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.032794 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.032803 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.135404 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.135453 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.135470 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.135489 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.135507 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.238607 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.238685 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.238705 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.238735 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.238759 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.270521 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.270578 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.270595 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.270620 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.270638 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: E1201 21:35:56.293365 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:56Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.299836 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.299947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.299966 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.299990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.300033 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: E1201 21:35:56.320927 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:56Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.326889 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.326940 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.326959 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.326984 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.327004 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: E1201 21:35:56.349196 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:56Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.354330 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.354373 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.354392 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.354416 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.354434 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: E1201 21:35:56.373879 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:56Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.379445 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.379508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.379525 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.379552 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.379572 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: E1201 21:35:56.400404 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:35:56Z is after 2025-08-24T17:21:41Z" Dec 01 21:35:56 crc kubenswrapper[4857]: E1201 21:35:56.400811 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.403196 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.403284 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.403307 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.403332 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.403471 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.506147 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.506202 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.506219 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.506243 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.506259 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.609224 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.609285 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.609304 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.609332 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.609350 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.711675 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.711749 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.711766 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.711792 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.711809 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.814674 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.814738 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.814760 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.814789 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.814811 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.834411 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.834416 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:56 crc kubenswrapper[4857]: E1201 21:35:56.834566 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:56 crc kubenswrapper[4857]: E1201 21:35:56.834745 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.917898 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.917944 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.917962 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.917987 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:56 crc kubenswrapper[4857]: I1201 21:35:56.918004 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:56Z","lastTransitionTime":"2025-12-01T21:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.021558 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.021610 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.021627 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.021649 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.021665 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:57Z","lastTransitionTime":"2025-12-01T21:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.125364 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.125424 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.125446 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.125474 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.125495 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:57Z","lastTransitionTime":"2025-12-01T21:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.227825 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.227888 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.227909 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.227939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.227963 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:57Z","lastTransitionTime":"2025-12-01T21:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.331436 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.331485 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.331504 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.331529 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.331546 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:57Z","lastTransitionTime":"2025-12-01T21:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.434271 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.434306 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.434317 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.434337 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.434349 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:57Z","lastTransitionTime":"2025-12-01T21:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.537673 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.537704 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.537717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.537734 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.537747 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:57Z","lastTransitionTime":"2025-12-01T21:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.641692 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.641759 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.641780 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.641813 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.641837 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:57Z","lastTransitionTime":"2025-12-01T21:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.744939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.744984 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.744997 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.745013 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.745026 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:57Z","lastTransitionTime":"2025-12-01T21:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.834649 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:57 crc kubenswrapper[4857]: E1201 21:35:57.834835 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.834910 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:57 crc kubenswrapper[4857]: E1201 21:35:57.835142 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.847591 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.847658 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.847682 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.847710 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.847728 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:57Z","lastTransitionTime":"2025-12-01T21:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.950557 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.950604 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.950616 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.950634 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:57 crc kubenswrapper[4857]: I1201 21:35:57.950647 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:57Z","lastTransitionTime":"2025-12-01T21:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.053473 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.053533 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.053549 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.053571 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.053588 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:58Z","lastTransitionTime":"2025-12-01T21:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.157807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.157876 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.157898 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.157950 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.157975 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:58Z","lastTransitionTime":"2025-12-01T21:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.261226 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.261293 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.261312 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.261337 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.261355 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:58Z","lastTransitionTime":"2025-12-01T21:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.363606 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.363907 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.363999 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.364151 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.364268 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:58Z","lastTransitionTime":"2025-12-01T21:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.467947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.468024 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.468092 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.468116 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.468135 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:58Z","lastTransitionTime":"2025-12-01T21:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.571540 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.571802 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.571833 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.571858 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.571879 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:58Z","lastTransitionTime":"2025-12-01T21:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.675720 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.675782 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.675806 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.675836 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.675857 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:58Z","lastTransitionTime":"2025-12-01T21:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.779462 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.779510 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.779529 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.779552 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.779570 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:58Z","lastTransitionTime":"2025-12-01T21:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.834596 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.834695 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:35:58 crc kubenswrapper[4857]: E1201 21:35:58.834758 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:35:58 crc kubenswrapper[4857]: E1201 21:35:58.836140 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.882536 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.882592 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.882611 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.882635 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.882654 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:58Z","lastTransitionTime":"2025-12-01T21:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.985817 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.985895 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.985915 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.985949 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:58 crc kubenswrapper[4857]: I1201 21:35:58.985973 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:58Z","lastTransitionTime":"2025-12-01T21:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.089372 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.089450 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.089476 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.089508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.089534 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:59Z","lastTransitionTime":"2025-12-01T21:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.192892 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.192961 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.192989 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.193021 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.193083 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:59Z","lastTransitionTime":"2025-12-01T21:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.296678 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.296746 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.296765 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.296793 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.296813 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:59Z","lastTransitionTime":"2025-12-01T21:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.399506 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.399589 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.399612 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.399642 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.399666 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:59Z","lastTransitionTime":"2025-12-01T21:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.502919 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.502979 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.502998 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.503022 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.503078 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:59Z","lastTransitionTime":"2025-12-01T21:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.605853 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.605918 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.605942 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.605970 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.605990 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:59Z","lastTransitionTime":"2025-12-01T21:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.709747 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.709800 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.709816 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.709842 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.709860 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:59Z","lastTransitionTime":"2025-12-01T21:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.812538 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.812600 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.812618 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.812643 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.812660 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:59Z","lastTransitionTime":"2025-12-01T21:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.834659 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.834763 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:35:59 crc kubenswrapper[4857]: E1201 21:35:59.834790 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:35:59 crc kubenswrapper[4857]: E1201 21:35:59.834937 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.915155 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.915311 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.915332 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.915356 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:35:59 crc kubenswrapper[4857]: I1201 21:35:59.915376 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:35:59Z","lastTransitionTime":"2025-12-01T21:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.018386 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.018451 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.018476 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.018507 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.018531 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:00Z","lastTransitionTime":"2025-12-01T21:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.121718 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.121802 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.121828 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.121860 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.121877 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:00Z","lastTransitionTime":"2025-12-01T21:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.225076 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.225119 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.225135 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.225159 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.225178 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:00Z","lastTransitionTime":"2025-12-01T21:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.328532 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.328595 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.328626 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.328662 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.328688 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:00Z","lastTransitionTime":"2025-12-01T21:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.432309 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.432362 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.432386 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.432416 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.432439 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:00Z","lastTransitionTime":"2025-12-01T21:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.535237 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.535292 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.535309 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.535333 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.535350 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:00Z","lastTransitionTime":"2025-12-01T21:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.638324 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.638375 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.638393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.638421 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.638443 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:00Z","lastTransitionTime":"2025-12-01T21:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.741490 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.741542 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.741554 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.741575 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.741588 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:00Z","lastTransitionTime":"2025-12-01T21:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.834598 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.834692 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:00 crc kubenswrapper[4857]: E1201 21:36:00.834828 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:00 crc kubenswrapper[4857]: E1201 21:36:00.835000 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.845028 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.845115 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.845134 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.845160 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.845180 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:00Z","lastTransitionTime":"2025-12-01T21:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.948377 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.948419 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.948431 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.948448 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:00 crc kubenswrapper[4857]: I1201 21:36:00.948460 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:00Z","lastTransitionTime":"2025-12-01T21:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.051735 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.051807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.051828 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.051856 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.051876 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:01Z","lastTransitionTime":"2025-12-01T21:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.154269 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.154325 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.154338 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.154362 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.154376 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:01Z","lastTransitionTime":"2025-12-01T21:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.258089 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.258132 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.258171 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.258194 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.258208 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:01Z","lastTransitionTime":"2025-12-01T21:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.361478 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.361532 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.361546 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.361568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.361582 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:01Z","lastTransitionTime":"2025-12-01T21:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.464852 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.465001 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.465013 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.465084 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.465100 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:01Z","lastTransitionTime":"2025-12-01T21:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.568804 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.568889 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.568903 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.568956 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.568973 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:01Z","lastTransitionTime":"2025-12-01T21:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.673649 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.673741 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.673785 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.673819 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.673838 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:01Z","lastTransitionTime":"2025-12-01T21:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.777380 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.777446 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.777530 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.777568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.777590 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:01Z","lastTransitionTime":"2025-12-01T21:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.834664 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.834735 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:01 crc kubenswrapper[4857]: E1201 21:36:01.834906 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:01 crc kubenswrapper[4857]: E1201 21:36:01.835173 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.855467 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.877467 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.882749 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.882801 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.882821 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.882852 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.882872 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:01Z","lastTransitionTime":"2025-12-01T21:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.898910 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.916134 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.937937 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.972144 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:33Z\\\",\\\"message\\\":\\\" 21:35:33.310121 6430 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:33.310216 6430 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310592 6430 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:33.310642 6430 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310687 6430 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.311649 6430 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:33.311689 6430 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:33.311774 6430 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:33.311807 6430 factory.go:656] Stopping watch factory\\\\nI1201 21:35:33.311829 6430 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:33.311881 6430 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:33.311899 6430 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.986791 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.986867 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.986890 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.986917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.986932 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:01Z","lastTransitionTime":"2025-12-01T21:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:01 crc kubenswrapper[4857]: I1201 21:36:01.987675 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:01Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.002060 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.018931 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.037532 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116aa470-b443-48c1-91ce-bf63e232b25f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://005b5e00ddaba85767652a371cfdae9b98238fc5b1ec89beecdb2c18d9b2062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.059106 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.074409 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.090656 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.091132 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.091162 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.091193 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.091216 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.091230 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:02Z","lastTransitionTime":"2025-12-01T21:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.111810 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.132971 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.158704 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.203162 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.203250 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.203262 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.203280 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.203291 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:02Z","lastTransitionTime":"2025-12-01T21:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.206902 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e38c5755141156082b254a121520d58f44c5cef43ad63c3cadb2f1cf8e186e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:53Z\\\",\\\"message\\\":\\\"2025-12-01T21:35:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bba25fa0-e0dc-4aea-9563-b9ea666147ec\\\\n2025-12-01T21:35:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bba25fa0-e0dc-4aea-9563-b9ea666147ec to /host/opt/cni/bin/\\\\n2025-12-01T21:35:08Z [verbose] multus-daemon started\\\\n2025-12-01T21:35:08Z [verbose] Readiness Indicator file check\\\\n2025-12-01T21:35:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.221918 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e231d4d6-c755-4e41-bfe2-d65bd9765357\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.306730 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.306804 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.306827 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.306858 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.306879 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:02Z","lastTransitionTime":"2025-12-01T21:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.410891 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.410969 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.410990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.411024 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.411080 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:02Z","lastTransitionTime":"2025-12-01T21:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.513881 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.513948 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.513966 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.513993 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.514013 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:02Z","lastTransitionTime":"2025-12-01T21:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.616734 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.616819 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.616847 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.616877 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.616899 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:02Z","lastTransitionTime":"2025-12-01T21:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.720268 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.720346 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.720358 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.720381 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.720398 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:02Z","lastTransitionTime":"2025-12-01T21:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.824015 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.824134 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.824150 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.824173 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.824186 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:02Z","lastTransitionTime":"2025-12-01T21:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.834702 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.834883 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:02 crc kubenswrapper[4857]: E1201 21:36:02.835084 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:02 crc kubenswrapper[4857]: E1201 21:36:02.835465 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.835967 4857 scope.go:117] "RemoveContainer" containerID="7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.927767 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.927825 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.927840 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.927868 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:02 crc kubenswrapper[4857]: I1201 21:36:02.927886 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:02Z","lastTransitionTime":"2025-12-01T21:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.031828 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.032836 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.032947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.033066 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.033181 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:03Z","lastTransitionTime":"2025-12-01T21:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.135921 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.135948 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.135958 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.135975 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.135987 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:03Z","lastTransitionTime":"2025-12-01T21:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.238414 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.238467 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.238486 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.238511 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.238529 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:03Z","lastTransitionTime":"2025-12-01T21:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.342285 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.342338 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.342355 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.342380 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.342398 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:03Z","lastTransitionTime":"2025-12-01T21:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.376598 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/2.log" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.381355 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411"} Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.381916 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.402766 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.417523 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.435737 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.445518 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.445590 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.445612 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.445640 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.445659 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:03Z","lastTransitionTime":"2025-12-01T21:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.451496 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.470055 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.492095 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.508188 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.540710 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:33Z\\\",\\\"message\\\":\\\" 21:35:33.310121 6430 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:33.310216 6430 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310592 6430 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:33.310642 6430 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310687 6430 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.311649 6430 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:33.311689 6430 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:33.311774 6430 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:33.311807 6430 factory.go:656] Stopping watch factory\\\\nI1201 21:35:33.311829 6430 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:33.311881 6430 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:33.311899 6430 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:36:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.554225 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.554268 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.554278 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.554295 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.554306 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:03Z","lastTransitionTime":"2025-12-01T21:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.563027 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.583694 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.598677 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.619651 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.650485 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116aa470-b443-48c1-91ce-bf63e232b25f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://005b5e00ddaba85767652a371cfdae9b98238fc5b1ec89beecdb2c18d9b2062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.656513 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.656549 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.656560 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.656580 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.656593 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:03Z","lastTransitionTime":"2025-12-01T21:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.669187 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.695810 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e38c5755141156082b254a121520d58f44c5cef43ad63c3cadb2f1cf8e186e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:53Z\\\",\\\"message\\\":\\\"2025-12-01T21:35:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bba25fa0-e0dc-4aea-9563-b9ea666147ec\\\\n2025-12-01T21:35:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bba25fa0-e0dc-4aea-9563-b9ea666147ec to /host/opt/cni/bin/\\\\n2025-12-01T21:35:08Z [verbose] multus-daemon started\\\\n2025-12-01T21:35:08Z [verbose] Readiness Indicator file check\\\\n2025-12-01T21:35:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.713638 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e231d4d6-c755-4e41-bfe2-d65bd9765357\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.730078 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.745169 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:03Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.759563 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.759629 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.759645 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.759676 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.759693 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:03Z","lastTransitionTime":"2025-12-01T21:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.834350 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:03 crc kubenswrapper[4857]: E1201 21:36:03.834511 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.834368 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:03 crc kubenswrapper[4857]: E1201 21:36:03.834755 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.861682 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.861764 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.861779 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.861805 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.861820 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:03Z","lastTransitionTime":"2025-12-01T21:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.964700 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.964771 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.964795 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.964833 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:03 crc kubenswrapper[4857]: I1201 21:36:03.964860 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:03Z","lastTransitionTime":"2025-12-01T21:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.067565 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.067617 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.067631 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.067653 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.067670 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:04Z","lastTransitionTime":"2025-12-01T21:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.113130 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.113354 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:08.113325537 +0000 UTC m=+146.603387854 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.170080 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.170144 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.170167 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.170196 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.170218 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:04Z","lastTransitionTime":"2025-12-01T21:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.214543 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.214599 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.214632 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.214695 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.214828 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.214834 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.214852 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.214891 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.214901 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:37:08.214879632 +0000 UTC m=+146.704941969 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.214911 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.214941 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 21:37:08.214915813 +0000 UTC m=+146.704978170 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.214992 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 21:37:08.214958784 +0000 UTC m=+146.705021141 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.215020 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.215138 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.215175 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.215311 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 21:37:08.215266832 +0000 UTC m=+146.705329289 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.272666 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.272950 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.273065 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.273154 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.273239 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:04Z","lastTransitionTime":"2025-12-01T21:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.377095 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.377162 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.377185 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.377215 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.377234 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:04Z","lastTransitionTime":"2025-12-01T21:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.481264 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.481337 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.481367 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.481401 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.481425 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:04Z","lastTransitionTime":"2025-12-01T21:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.584669 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.585028 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.585367 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.585493 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.585515 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:04Z","lastTransitionTime":"2025-12-01T21:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.689441 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.689491 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.689502 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.689520 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.689531 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:04Z","lastTransitionTime":"2025-12-01T21:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.791799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.791853 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.791867 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.791885 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.791899 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:04Z","lastTransitionTime":"2025-12-01T21:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.834707 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.834718 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.834899 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:04 crc kubenswrapper[4857]: E1201 21:36:04.834966 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.895212 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.895257 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.895268 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.895283 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.895292 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:04Z","lastTransitionTime":"2025-12-01T21:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.998322 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.998366 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.998376 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.998394 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:04 crc kubenswrapper[4857]: I1201 21:36:04.998410 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:04Z","lastTransitionTime":"2025-12-01T21:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.101857 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.101932 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.101943 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.101965 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.101975 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:05Z","lastTransitionTime":"2025-12-01T21:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.204685 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.204739 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.204754 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.204778 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.204792 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:05Z","lastTransitionTime":"2025-12-01T21:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.307968 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.308081 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.308103 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.308132 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.308151 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:05Z","lastTransitionTime":"2025-12-01T21:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.393572 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/3.log" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.394781 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/2.log" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.398162 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411" exitCode=1 Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.398214 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411"} Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.398284 4857 scope.go:117] "RemoveContainer" containerID="7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.399587 4857 scope.go:117] "RemoveContainer" containerID="9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411" Dec 01 21:36:05 crc kubenswrapper[4857]: E1201 21:36:05.399847 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.411449 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.411481 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.411492 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.411511 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.411523 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:05Z","lastTransitionTime":"2025-12-01T21:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.421161 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.444725 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.462458 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.479763 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.503207 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.514813 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.514859 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.514876 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.514905 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.514923 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:05Z","lastTransitionTime":"2025-12-01T21:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.519597 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.548016 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:33Z\\\",\\\"message\\\":\\\" 21:35:33.310121 6430 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:33.310216 6430 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310592 6430 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:33.310642 6430 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310687 6430 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.311649 6430 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:33.311689 6430 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:33.311774 6430 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:33.311807 6430 factory.go:656] Stopping watch factory\\\\nI1201 21:35:33.311829 6430 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:33.311881 6430 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:33.311899 6430 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:36:04Z\\\",\\\"message\\\":\\\"dressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1201 21:36:04.263477 6825 services_controller.go:453] Built service openshift-machine-api/control-plane-machine-set-operator template LB for network=default: []services.LB{}\\\\nI1201 21:36:04.263536 6825 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"1dc899db-4498-4b7a-8437-861940b962e7\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:36:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.563354 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.583385 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.605263 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.618849 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.618936 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.618964 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.619000 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.619026 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:05Z","lastTransitionTime":"2025-12-01T21:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.622195 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116aa470-b443-48c1-91ce-bf63e232b25f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://005b5e00ddaba85767652a371cfdae9b98238fc5b1ec89beecdb2c18d9b2062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.644879 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.666230 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.678326 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e231d4d6-c755-4e41-bfe2-d65bd9765357\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.690541 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.704904 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.718868 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.721490 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.721533 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.721543 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.721560 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.721572 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:05Z","lastTransitionTime":"2025-12-01T21:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.733475 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e38c5755141156082b254a121520d58f44c5cef43ad63c3cadb2f1cf8e186e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:53Z\\\",\\\"message\\\":\\\"2025-12-01T21:35:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bba25fa0-e0dc-4aea-9563-b9ea666147ec\\\\n2025-12-01T21:35:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bba25fa0-e0dc-4aea-9563-b9ea666147ec to /host/opt/cni/bin/\\\\n2025-12-01T21:35:08Z [verbose] multus-daemon started\\\\n2025-12-01T21:35:08Z [verbose] Readiness Indicator file check\\\\n2025-12-01T21:35:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.825122 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.825203 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.825228 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.825261 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.825284 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:05Z","lastTransitionTime":"2025-12-01T21:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.833964 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.834028 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:05 crc kubenswrapper[4857]: E1201 21:36:05.834160 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:05 crc kubenswrapper[4857]: E1201 21:36:05.834468 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.928264 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.928319 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.928333 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.928351 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:05 crc kubenswrapper[4857]: I1201 21:36:05.928365 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:05Z","lastTransitionTime":"2025-12-01T21:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.031734 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.031822 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.031846 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.031885 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.031911 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.135800 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.135876 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.135894 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.135926 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.135944 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.238847 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.238917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.238935 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.238964 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.238979 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.341475 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.341525 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.341543 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.341565 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.341599 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.404461 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/3.log" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.445185 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.445237 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.445255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.445278 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.445294 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.548261 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.548299 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.548309 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.548326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.548347 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.651671 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.651714 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.651723 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.651741 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.651757 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.729171 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.729221 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.729231 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.729248 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.729259 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: E1201 21:36:06.743463 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.749174 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.749223 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.749234 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.749249 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.749261 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: E1201 21:36:06.762797 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.766739 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.766790 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.766806 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.766825 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.766838 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: E1201 21:36:06.780032 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.784075 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.784121 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.784133 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.784155 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.784166 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: E1201 21:36:06.805992 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.810395 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.810440 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.810459 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.810484 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.810503 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: E1201 21:36:06.829954 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T21:36:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea7fffe2-6d6d-498a-b750-d72f9c51be69\\\",\\\"systemUUID\\\":\\\"0d498ed1-bf21-4e30-b0f7-bd8e4f32b853\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:06 crc kubenswrapper[4857]: E1201 21:36:06.830114 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.831648 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.831674 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.831682 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.831696 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.831705 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.833998 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.834024 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:06 crc kubenswrapper[4857]: E1201 21:36:06.834112 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:06 crc kubenswrapper[4857]: E1201 21:36:06.834197 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.934657 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.934705 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.934720 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.934738 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:06 crc kubenswrapper[4857]: I1201 21:36:06.934750 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:06Z","lastTransitionTime":"2025-12-01T21:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.038162 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.038236 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.038255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.038282 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.038301 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:07Z","lastTransitionTime":"2025-12-01T21:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.141502 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.141564 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.141581 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.141610 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.141628 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:07Z","lastTransitionTime":"2025-12-01T21:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.245218 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.245266 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.245278 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.245297 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.245312 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:07Z","lastTransitionTime":"2025-12-01T21:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.348210 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.348283 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.348305 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.348330 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.348349 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:07Z","lastTransitionTime":"2025-12-01T21:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.450834 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.450890 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.450901 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.450919 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.450930 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:07Z","lastTransitionTime":"2025-12-01T21:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.554278 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.554347 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.554373 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.554401 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.554419 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:07Z","lastTransitionTime":"2025-12-01T21:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.657472 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.657549 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.657571 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.657605 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.657627 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:07Z","lastTransitionTime":"2025-12-01T21:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.760165 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.760271 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.760296 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.760332 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.760356 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:07Z","lastTransitionTime":"2025-12-01T21:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.834113 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.834123 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:07 crc kubenswrapper[4857]: E1201 21:36:07.834254 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:07 crc kubenswrapper[4857]: E1201 21:36:07.834382 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.863304 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.863374 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.863393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.863418 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.863435 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:07Z","lastTransitionTime":"2025-12-01T21:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.972782 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.972837 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.972854 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.972876 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:07 crc kubenswrapper[4857]: I1201 21:36:07.972896 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:07Z","lastTransitionTime":"2025-12-01T21:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.075520 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.075623 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.075643 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.075669 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.075688 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:08Z","lastTransitionTime":"2025-12-01T21:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.178952 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.179011 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.179030 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.179081 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.179100 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:08Z","lastTransitionTime":"2025-12-01T21:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.282286 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.282360 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.282385 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.282416 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.282439 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:08Z","lastTransitionTime":"2025-12-01T21:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.385417 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.385533 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.385551 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.385574 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.385591 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:08Z","lastTransitionTime":"2025-12-01T21:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.488292 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.488356 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.488378 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.488407 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.488432 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:08Z","lastTransitionTime":"2025-12-01T21:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.590614 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.590677 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.590703 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.590735 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.590759 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:08Z","lastTransitionTime":"2025-12-01T21:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.693906 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.693981 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.694000 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.694025 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.694073 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:08Z","lastTransitionTime":"2025-12-01T21:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.797511 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.797565 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.797579 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.797600 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.797615 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:08Z","lastTransitionTime":"2025-12-01T21:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.834247 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.834353 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:08 crc kubenswrapper[4857]: E1201 21:36:08.834449 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:08 crc kubenswrapper[4857]: E1201 21:36:08.834570 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.901034 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.901131 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.901148 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.901195 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:08 crc kubenswrapper[4857]: I1201 21:36:08.901216 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:08Z","lastTransitionTime":"2025-12-01T21:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.004490 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.004568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.004593 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.004626 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.004644 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:09Z","lastTransitionTime":"2025-12-01T21:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.108185 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.108249 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.108267 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.108293 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.108315 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:09Z","lastTransitionTime":"2025-12-01T21:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.212317 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.212413 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.212444 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.212473 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.212502 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:09Z","lastTransitionTime":"2025-12-01T21:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.315839 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.315905 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.315926 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.315951 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.315972 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:09Z","lastTransitionTime":"2025-12-01T21:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.418425 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.418481 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.418499 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.418522 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.418539 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:09Z","lastTransitionTime":"2025-12-01T21:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.521184 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.521260 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.521283 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.521313 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.521334 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:09Z","lastTransitionTime":"2025-12-01T21:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.624979 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.625035 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.625104 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.625134 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.625156 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:09Z","lastTransitionTime":"2025-12-01T21:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.728672 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.728740 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.728757 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.728783 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.728800 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:09Z","lastTransitionTime":"2025-12-01T21:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.832518 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.832587 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.832604 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.832628 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.832645 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:09Z","lastTransitionTime":"2025-12-01T21:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.834076 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.834132 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:09 crc kubenswrapper[4857]: E1201 21:36:09.834227 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:09 crc kubenswrapper[4857]: E1201 21:36:09.834485 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.935862 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.935911 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.935932 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.935952 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:09 crc kubenswrapper[4857]: I1201 21:36:09.935969 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:09Z","lastTransitionTime":"2025-12-01T21:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.038980 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.039051 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.039075 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.039137 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.039158 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:10Z","lastTransitionTime":"2025-12-01T21:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.142638 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.142703 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.142726 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.142755 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.142776 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:10Z","lastTransitionTime":"2025-12-01T21:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.246905 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.247336 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.247358 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.247387 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.247425 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:10Z","lastTransitionTime":"2025-12-01T21:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.350643 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.350729 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.350767 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.350797 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.350818 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:10Z","lastTransitionTime":"2025-12-01T21:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.455645 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.455724 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.455750 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.455781 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.455805 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:10Z","lastTransitionTime":"2025-12-01T21:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.559373 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.559453 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.559473 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.559501 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.559522 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:10Z","lastTransitionTime":"2025-12-01T21:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.663434 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.663545 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.663566 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.663590 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.663655 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:10Z","lastTransitionTime":"2025-12-01T21:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.767726 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.767798 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.767818 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.767843 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.767862 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:10Z","lastTransitionTime":"2025-12-01T21:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.834163 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:10 crc kubenswrapper[4857]: E1201 21:36:10.834357 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.834169 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:10 crc kubenswrapper[4857]: E1201 21:36:10.834463 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.872215 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.872263 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.872277 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.872300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.872326 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:10Z","lastTransitionTime":"2025-12-01T21:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.975168 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.975230 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.975251 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.975283 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:10 crc kubenswrapper[4857]: I1201 21:36:10.975307 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:10Z","lastTransitionTime":"2025-12-01T21:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.078642 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.078699 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.078711 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.078736 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.078751 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:11Z","lastTransitionTime":"2025-12-01T21:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.182486 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.182644 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.182666 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.182729 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.182751 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:11Z","lastTransitionTime":"2025-12-01T21:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.286022 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.286174 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.286195 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.286225 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.286245 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:11Z","lastTransitionTime":"2025-12-01T21:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.389361 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.389409 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.389421 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.389441 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.389455 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:11Z","lastTransitionTime":"2025-12-01T21:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.493390 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.493491 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.493517 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.493586 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.493612 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:11Z","lastTransitionTime":"2025-12-01T21:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.596799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.596901 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.596928 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.596965 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.596995 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:11Z","lastTransitionTime":"2025-12-01T21:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.699553 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.699622 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.699639 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.699668 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.699689 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:11Z","lastTransitionTime":"2025-12-01T21:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.801973 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.802152 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.802180 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.802213 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.802238 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:11Z","lastTransitionTime":"2025-12-01T21:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.834773 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:11 crc kubenswrapper[4857]: E1201 21:36:11.835067 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.835107 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:11 crc kubenswrapper[4857]: E1201 21:36:11.836243 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.853960 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e231d4d6-c755-4e41-bfe2-d65bd9765357\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d70e8145baa78e5bf370b1bc64ecea828f2aa5e41462a2ccb8d22946a8ebb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55db99795e469f1ea27473bebba80f33aa7f8ea12384f17d0e72cf6fcffe1b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36239dbca41d05afe26ec2a3e79749ca779eb16964cac53fa5190653ec20a188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a34f46dfad026392043a540af8da117173dfcfb445b6b8066abb17968bf394\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.869582 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa27b00197c8fdc76017295247168cf0e85470d5273a18801eac80de30a99b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.889417 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268283eac88bf62b7ce14d7998a4dec223108f011e6f3699d225b519db5f572c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.905113 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.905193 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.905218 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.905249 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.905273 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:11Z","lastTransitionTime":"2025-12-01T21:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.917656 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tczkh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6721c7e7-7840-4241-a3b2-d988c6dcf215\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94980db7c14852ffba377684d1a1e996ecf9625fcce5a716227711083020f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f0c97c8ad85f136ad289dd2de717e89a90d17f225545b6b1f9b5f9913773703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ca2d29af4b629583d443705e7c6d6c407564c4407bdbbd1ed4e7c8a2a7819ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0652595fddc4e82503f1d4595f4c3c74995fa65d67eef8332e4906130888b0bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://492059746b076a4593fdce734e2bbd94430270cb157f7313011d54c8cb6e93ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79978e17c19118723ed8e5cbb2a7b11ea466428a662bf910afa169ca8218491f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d59deb2860e4f8b71869aaea325d52595a9716bf02a8fa253a33da93f37e31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4m47\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tczkh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.935597 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dfxtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc7d55e5-8085-40b7-8147-8508ee6c3f61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e38c5755141156082b254a121520d58f44c5cef43ad63c3cadb2f1cf8e186e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:53Z\\\",\\\"message\\\":\\\"2025-12-01T21:35:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bba25fa0-e0dc-4aea-9563-b9ea666147ec\\\\n2025-12-01T21:35:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bba25fa0-e0dc-4aea-9563-b9ea666147ec to /host/opt/cni/bin/\\\\n2025-12-01T21:35:08Z [verbose] multus-daemon started\\\\n2025-12-01T21:35:08Z [verbose] Readiness Indicator file check\\\\n2025-12-01T21:35:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97tnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dfxtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.957028 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef95cff-01c5-4ef5-8c68-bf2fcb13e17d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655a546c8603ac0e0198bdc08cc006809fe7920b0bf1861f3c2990a4e7f7ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fec40d794952aabd446ff1bcc36c81e2e666c9d47422b1abe3c48e04e1f101\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3ff8383efc76def8b1110aac9fcb1bed1686a45072ece7ab2a5b2dc702d7c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.974098 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:11 crc kubenswrapper[4857]: I1201 21:36:11.995612 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.008410 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.008514 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.008533 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.008558 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.008576 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:12Z","lastTransitionTime":"2025-12-01T21:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.014723 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac5dfac0-8933-4db8-9cb6-5dca0537d066\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72d680d73faf119b13c1008dd347a2913ee132e7c4c0d11c0cfda39118e97c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bqvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-89689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.031694 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ss5tk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n2fsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.051036 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.082169 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"557fd3cd-9ab9-4135-af14-f748fdc34a90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c6b24271c27b0fe9a9cb29d3973abd159eb2886312e053c8c8b2f46ac1b8692\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:35:33Z\\\",\\\"message\\\":\\\" 21:35:33.310121 6430 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 21:35:33.310216 6430 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310592 6430 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 21:35:33.310642 6430 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.310687 6430 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 21:35:33.311649 6430 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 21:35:33.311689 6430 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 21:35:33.311774 6430 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 21:35:33.311807 6430 factory.go:656] Stopping watch factory\\\\nI1201 21:35:33.311829 6430 ovnkube.go:599] Stopped ovnkube\\\\nI1201 21:35:33.311881 6430 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 21:35:33.311899 6430 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 21:35:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T21:36:04Z\\\",\\\"message\\\":\\\"dressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1201 21:36:04.263477 6825 services_controller.go:453] Built service openshift-machine-api/control-plane-machine-set-operator template LB for network=default: []services.LB{}\\\\nI1201 21:36:04.263536 6825 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"1dc899db-4498-4b7a-8437-861940b962e7\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T21:36:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:35:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftp2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mdc5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.099461 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-49cdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"112193dd-6798-4528-890e-a9713a21b3cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69d3a9dd809854efc3bc45197cd89bb9335bd05349214fc83b0260a7f17c663d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4g49r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-49cdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.112910 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.112969 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.112987 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.113150 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.113172 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:12Z","lastTransitionTime":"2025-12-01T21:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.119693 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8e9109-4f8d-4410-a463-368cb7cfbee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01adb3cb22c95ca7a0cfc818a150e8f1612bb8b885715a1e50f208a79240875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f419f63c5727d35ee9ab53ac65a0e53291efa9c6367dd239a885fe59fa3a237d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7498\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c8mcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.144013 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d25a23a8-5aa2-4324-8d55-e8000ac37979\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.161760 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116aa470-b443-48c1-91ce-bf63e232b25f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://005b5e00ddaba85767652a371cfdae9b98238fc5b1ec89beecdb2c18d9b2062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67fbc3d8203e54eb889ba9b6f453a893ac9b52276e00d8ef8a50924c7518c192\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T21:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T21:34:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:34:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.183443 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23208bcb9655da38a6a937181392654d6f547d5ea87bb829f3fcc0e4a9483e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966ce35e310b42ea53781ea8f7fc933d61e9543043cfee4cc13e0e82b78febed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.200685 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7x8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f63b0e62-e9e8-432e-9a1b-f5be926fc2ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T21:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b670133252a49c67e370f585ae3ad3440e16d16eab7dcfa3015c1088a59fd1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T21:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2pv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T21:35:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7x8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T21:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.216247 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.216302 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.216319 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.216345 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.216364 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:12Z","lastTransitionTime":"2025-12-01T21:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.319723 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.319788 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.319807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.319835 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.319853 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:12Z","lastTransitionTime":"2025-12-01T21:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.422808 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.422861 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.422873 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.422889 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.422905 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:12Z","lastTransitionTime":"2025-12-01T21:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.525874 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.525946 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.525968 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.525995 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.526018 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:12Z","lastTransitionTime":"2025-12-01T21:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.629156 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.629264 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.629282 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.629309 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.629330 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:12Z","lastTransitionTime":"2025-12-01T21:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.733147 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.733203 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.733220 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.733244 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.733262 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:12Z","lastTransitionTime":"2025-12-01T21:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.833976 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.834176 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:12 crc kubenswrapper[4857]: E1201 21:36:12.834425 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:12 crc kubenswrapper[4857]: E1201 21:36:12.834605 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.835850 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.835901 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.835923 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.835954 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.835977 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:12Z","lastTransitionTime":"2025-12-01T21:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.939695 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.939770 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.939795 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.939824 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:12 crc kubenswrapper[4857]: I1201 21:36:12.939845 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:12Z","lastTransitionTime":"2025-12-01T21:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.042931 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.043006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.043034 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.043102 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.043131 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:13Z","lastTransitionTime":"2025-12-01T21:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.146912 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.146960 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.146972 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.146990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.147004 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:13Z","lastTransitionTime":"2025-12-01T21:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.250331 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.250365 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.250379 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.250402 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.250415 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:13Z","lastTransitionTime":"2025-12-01T21:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.353945 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.353987 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.353997 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.354013 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.354022 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:13Z","lastTransitionTime":"2025-12-01T21:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.456564 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.456627 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.456644 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.456669 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.456689 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:13Z","lastTransitionTime":"2025-12-01T21:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.559692 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.559761 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.559774 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.559793 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.559805 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:13Z","lastTransitionTime":"2025-12-01T21:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.662362 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.662418 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.662436 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.662459 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.662477 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:13Z","lastTransitionTime":"2025-12-01T21:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.765102 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.765151 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.765162 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.765181 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.765194 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:13Z","lastTransitionTime":"2025-12-01T21:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.834565 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.834616 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:13 crc kubenswrapper[4857]: E1201 21:36:13.834741 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:13 crc kubenswrapper[4857]: E1201 21:36:13.834911 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.867479 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.867524 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.867535 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.867551 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.867571 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:13Z","lastTransitionTime":"2025-12-01T21:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.971343 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.971385 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.971394 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.971409 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:13 crc kubenswrapper[4857]: I1201 21:36:13.971419 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:13Z","lastTransitionTime":"2025-12-01T21:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.073734 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.073808 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.073827 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.073854 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.073872 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:14Z","lastTransitionTime":"2025-12-01T21:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.177754 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.177839 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.177858 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.177883 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.177927 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:14Z","lastTransitionTime":"2025-12-01T21:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.282104 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.282208 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.282230 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.282288 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.282309 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:14Z","lastTransitionTime":"2025-12-01T21:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.385305 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.385370 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.385389 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.385412 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.385431 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:14Z","lastTransitionTime":"2025-12-01T21:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.488516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.488603 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.488622 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.488674 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.488691 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:14Z","lastTransitionTime":"2025-12-01T21:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.591432 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.591488 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.591508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.591534 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.591550 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:14Z","lastTransitionTime":"2025-12-01T21:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.694489 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.694550 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.694567 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.694592 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.694609 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:14Z","lastTransitionTime":"2025-12-01T21:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.797421 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.797796 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.797950 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.798132 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.798293 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:14Z","lastTransitionTime":"2025-12-01T21:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.834288 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.834305 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:14 crc kubenswrapper[4857]: E1201 21:36:14.834422 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:14 crc kubenswrapper[4857]: E1201 21:36:14.834568 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.902264 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.902363 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.902385 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.902442 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:14 crc kubenswrapper[4857]: I1201 21:36:14.902463 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:14Z","lastTransitionTime":"2025-12-01T21:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.005135 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.005255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.005326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.005393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.005416 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:15Z","lastTransitionTime":"2025-12-01T21:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.108844 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.108897 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.108914 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.108938 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.108954 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:15Z","lastTransitionTime":"2025-12-01T21:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.212184 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.212258 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.212276 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.212302 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.212320 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:15Z","lastTransitionTime":"2025-12-01T21:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.315356 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.315426 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.315444 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.315468 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.315486 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:15Z","lastTransitionTime":"2025-12-01T21:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.418669 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.419034 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.419229 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.419379 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.419499 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:15Z","lastTransitionTime":"2025-12-01T21:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.522446 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.522516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.522544 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.522576 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.522598 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:15Z","lastTransitionTime":"2025-12-01T21:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.625571 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.625692 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.625787 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.625864 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.625883 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:15Z","lastTransitionTime":"2025-12-01T21:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.729362 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.729414 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.729438 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.729460 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.729485 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:15Z","lastTransitionTime":"2025-12-01T21:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.832136 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.832202 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.832226 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.832256 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.832282 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:15Z","lastTransitionTime":"2025-12-01T21:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.833966 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.834028 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:15 crc kubenswrapper[4857]: E1201 21:36:15.834259 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:15 crc kubenswrapper[4857]: E1201 21:36:15.834556 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.940999 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.941132 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.941158 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.941196 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:15 crc kubenswrapper[4857]: I1201 21:36:15.941232 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:15Z","lastTransitionTime":"2025-12-01T21:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.045244 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.045340 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.045361 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.045387 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.045406 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:16Z","lastTransitionTime":"2025-12-01T21:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.148330 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.148367 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.148375 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.148389 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.148397 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:16Z","lastTransitionTime":"2025-12-01T21:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.251172 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.251244 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.251263 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.251290 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.251308 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:16Z","lastTransitionTime":"2025-12-01T21:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.354506 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.354561 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.354584 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.354615 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.354641 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:16Z","lastTransitionTime":"2025-12-01T21:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.458662 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.458727 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.458743 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.458766 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.458787 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:16Z","lastTransitionTime":"2025-12-01T21:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.562098 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.562135 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.562144 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.562157 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.562169 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:16Z","lastTransitionTime":"2025-12-01T21:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.665276 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.665326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.665338 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.665357 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.665370 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:16Z","lastTransitionTime":"2025-12-01T21:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.768524 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.768571 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.768582 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.768599 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.768611 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:16Z","lastTransitionTime":"2025-12-01T21:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.833910 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.833950 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:16 crc kubenswrapper[4857]: E1201 21:36:16.834135 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:16 crc kubenswrapper[4857]: E1201 21:36:16.834260 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.871094 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.871139 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.871152 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.871198 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.871213 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:16Z","lastTransitionTime":"2025-12-01T21:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.974515 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.974568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.974597 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.974619 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:16 crc kubenswrapper[4857]: I1201 21:36:16.974635 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:16Z","lastTransitionTime":"2025-12-01T21:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.077924 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.077997 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.078012 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.078041 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.078074 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:17Z","lastTransitionTime":"2025-12-01T21:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.098291 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.098363 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.098376 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.098393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.098408 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T21:36:17Z","lastTransitionTime":"2025-12-01T21:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.172384 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99"] Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.172953 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.176092 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.176910 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.177260 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.177598 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.259806 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=75.259765046 podStartE2EDuration="1m15.259765046s" podCreationTimestamp="2025-12-01 21:35:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:17.238233656 +0000 UTC m=+95.728295993" watchObservedRunningTime="2025-12-01 21:36:17.259765046 +0000 UTC m=+95.749827373" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.269578 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/de49f270-5ff5-43f7-8602-5fcfed94694f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.269644 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de49f270-5ff5-43f7-8602-5fcfed94694f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.269713 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/de49f270-5ff5-43f7-8602-5fcfed94694f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.269754 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/de49f270-5ff5-43f7-8602-5fcfed94694f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.269971 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/de49f270-5ff5-43f7-8602-5fcfed94694f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.295238 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-89689" podStartSLOduration=77.295198261 podStartE2EDuration="1m17.295198261s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:17.294934204 +0000 UTC m=+95.784996551" watchObservedRunningTime="2025-12-01 21:36:17.295198261 +0000 UTC m=+95.785260618" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.356303 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-49cdz" podStartSLOduration=77.356283472 podStartE2EDuration="1m17.356283472s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:17.355728648 +0000 UTC m=+95.845790965" watchObservedRunningTime="2025-12-01 21:36:17.356283472 +0000 UTC m=+95.846345789" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.370878 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/de49f270-5ff5-43f7-8602-5fcfed94694f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.370928 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de49f270-5ff5-43f7-8602-5fcfed94694f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.370960 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/de49f270-5ff5-43f7-8602-5fcfed94694f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.370991 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/de49f270-5ff5-43f7-8602-5fcfed94694f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.371027 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/de49f270-5ff5-43f7-8602-5fcfed94694f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.371641 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/de49f270-5ff5-43f7-8602-5fcfed94694f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.372626 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/de49f270-5ff5-43f7-8602-5fcfed94694f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.372705 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/de49f270-5ff5-43f7-8602-5fcfed94694f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.385948 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de49f270-5ff5-43f7-8602-5fcfed94694f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.412689 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c8mcr" podStartSLOduration=76.412661743 podStartE2EDuration="1m16.412661743s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:17.393082383 +0000 UTC m=+95.883144720" watchObservedRunningTime="2025-12-01 21:36:17.412661743 +0000 UTC m=+95.902724060" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.432796 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/de49f270-5ff5-43f7-8602-5fcfed94694f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-b4x99\" (UID: \"de49f270-5ff5-43f7-8602-5fcfed94694f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.455187 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=79.455159729 podStartE2EDuration="1m19.455159729s" podCreationTimestamp="2025-12-01 21:34:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:17.454478762 +0000 UTC m=+95.944541079" watchObservedRunningTime="2025-12-01 21:36:17.455159729 +0000 UTC m=+95.945222066" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.455693 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-r7x8v" podStartSLOduration=77.455687452 podStartE2EDuration="1m17.455687452s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:17.421650073 +0000 UTC m=+95.911712400" watchObservedRunningTime="2025-12-01 21:36:17.455687452 +0000 UTC m=+95.945749769" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.469176 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=30.469142396 podStartE2EDuration="30.469142396s" podCreationTimestamp="2025-12-01 21:35:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:17.468801328 +0000 UTC m=+95.958863645" watchObservedRunningTime="2025-12-01 21:36:17.469142396 +0000 UTC m=+95.959204713" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.492037 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.509958 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-dfxtl" podStartSLOduration=77.509932929 podStartE2EDuration="1m17.509932929s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:17.508882722 +0000 UTC m=+95.998945049" watchObservedRunningTime="2025-12-01 21:36:17.509932929 +0000 UTC m=+95.999995246" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.529173 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=48.52915064 podStartE2EDuration="48.52915064s" podCreationTimestamp="2025-12-01 21:35:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:17.528142944 +0000 UTC m=+96.018205271" watchObservedRunningTime="2025-12-01 21:36:17.52915064 +0000 UTC m=+96.019212967" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.835353 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:17 crc kubenswrapper[4857]: I1201 21:36:17.835465 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:17 crc kubenswrapper[4857]: E1201 21:36:17.835531 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:17 crc kubenswrapper[4857]: E1201 21:36:17.835716 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:18 crc kubenswrapper[4857]: I1201 21:36:18.457196 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" event={"ID":"de49f270-5ff5-43f7-8602-5fcfed94694f","Type":"ContainerStarted","Data":"4d8e87663d70e0b6192e036a3b65e69276a74f47c83a26b94bba1e0f68b465e5"} Dec 01 21:36:18 crc kubenswrapper[4857]: I1201 21:36:18.457260 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" event={"ID":"de49f270-5ff5-43f7-8602-5fcfed94694f","Type":"ContainerStarted","Data":"87d06780824a85c6e2bd1a5766c1796c2346c58586b8b4032f0d7447d9512bf5"} Dec 01 21:36:18 crc kubenswrapper[4857]: I1201 21:36:18.484932 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-tczkh" podStartSLOduration=78.484913834 podStartE2EDuration="1m18.484913834s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:17.58707325 +0000 UTC m=+96.077135577" watchObservedRunningTime="2025-12-01 21:36:18.484913834 +0000 UTC m=+96.974976141" Dec 01 21:36:18 crc kubenswrapper[4857]: I1201 21:36:18.833985 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:18 crc kubenswrapper[4857]: I1201 21:36:18.834005 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:18 crc kubenswrapper[4857]: E1201 21:36:18.834228 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:18 crc kubenswrapper[4857]: E1201 21:36:18.834406 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:19 crc kubenswrapper[4857]: I1201 21:36:19.834175 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:19 crc kubenswrapper[4857]: I1201 21:36:19.834268 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:19 crc kubenswrapper[4857]: I1201 21:36:19.835300 4857 scope.go:117] "RemoveContainer" containerID="9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411" Dec 01 21:36:19 crc kubenswrapper[4857]: E1201 21:36:19.835440 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:19 crc kubenswrapper[4857]: E1201 21:36:19.835575 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" Dec 01 21:36:19 crc kubenswrapper[4857]: E1201 21:36:19.835452 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:19 crc kubenswrapper[4857]: I1201 21:36:19.880609 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b4x99" podStartSLOduration=79.88059216 podStartE2EDuration="1m19.88059216s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:18.484868383 +0000 UTC m=+96.974930760" watchObservedRunningTime="2025-12-01 21:36:19.88059216 +0000 UTC m=+98.370654477" Dec 01 21:36:19 crc kubenswrapper[4857]: I1201 21:36:19.901133 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:19 crc kubenswrapper[4857]: E1201 21:36:19.902476 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:36:19 crc kubenswrapper[4857]: E1201 21:36:19.902603 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs podName:485d0d95-8002-4bc3-8935-8b8f4ab5b2b3 nodeName:}" failed. No retries permitted until 2025-12-01 21:37:23.902572732 +0000 UTC m=+162.392635089 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs") pod "network-metrics-daemon-n2fsz" (UID: "485d0d95-8002-4bc3-8935-8b8f4ab5b2b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 21:36:20 crc kubenswrapper[4857]: I1201 21:36:20.834190 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:20 crc kubenswrapper[4857]: E1201 21:36:20.834356 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:20 crc kubenswrapper[4857]: I1201 21:36:20.835163 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:20 crc kubenswrapper[4857]: E1201 21:36:20.835342 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:21 crc kubenswrapper[4857]: I1201 21:36:21.834266 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:21 crc kubenswrapper[4857]: I1201 21:36:21.834350 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:21 crc kubenswrapper[4857]: E1201 21:36:21.836110 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:21 crc kubenswrapper[4857]: E1201 21:36:21.836328 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:22 crc kubenswrapper[4857]: I1201 21:36:22.833810 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:22 crc kubenswrapper[4857]: I1201 21:36:22.833924 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:22 crc kubenswrapper[4857]: E1201 21:36:22.833992 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:22 crc kubenswrapper[4857]: E1201 21:36:22.834111 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:23 crc kubenswrapper[4857]: I1201 21:36:23.834563 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:23 crc kubenswrapper[4857]: E1201 21:36:23.835413 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:23 crc kubenswrapper[4857]: I1201 21:36:23.834612 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:23 crc kubenswrapper[4857]: E1201 21:36:23.836368 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:24 crc kubenswrapper[4857]: I1201 21:36:24.834723 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:24 crc kubenswrapper[4857]: E1201 21:36:24.834884 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:24 crc kubenswrapper[4857]: I1201 21:36:24.834934 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:24 crc kubenswrapper[4857]: E1201 21:36:24.835797 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:24 crc kubenswrapper[4857]: I1201 21:36:24.856523 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 01 21:36:25 crc kubenswrapper[4857]: I1201 21:36:25.834236 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:25 crc kubenswrapper[4857]: I1201 21:36:25.834262 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:25 crc kubenswrapper[4857]: E1201 21:36:25.834397 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:25 crc kubenswrapper[4857]: E1201 21:36:25.834729 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:26 crc kubenswrapper[4857]: I1201 21:36:26.834321 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:26 crc kubenswrapper[4857]: I1201 21:36:26.834450 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:26 crc kubenswrapper[4857]: E1201 21:36:26.834521 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:26 crc kubenswrapper[4857]: E1201 21:36:26.834613 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:27 crc kubenswrapper[4857]: I1201 21:36:27.834574 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:27 crc kubenswrapper[4857]: E1201 21:36:27.834740 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:27 crc kubenswrapper[4857]: I1201 21:36:27.835019 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:27 crc kubenswrapper[4857]: E1201 21:36:27.835151 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:28 crc kubenswrapper[4857]: I1201 21:36:28.834714 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:28 crc kubenswrapper[4857]: E1201 21:36:28.834847 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:28 crc kubenswrapper[4857]: I1201 21:36:28.834925 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:28 crc kubenswrapper[4857]: E1201 21:36:28.835192 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:29 crc kubenswrapper[4857]: I1201 21:36:29.834390 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:29 crc kubenswrapper[4857]: E1201 21:36:29.834812 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:29 crc kubenswrapper[4857]: I1201 21:36:29.834403 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:29 crc kubenswrapper[4857]: E1201 21:36:29.835450 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:30 crc kubenswrapper[4857]: I1201 21:36:30.834376 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:30 crc kubenswrapper[4857]: I1201 21:36:30.834487 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:30 crc kubenswrapper[4857]: E1201 21:36:30.834547 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:30 crc kubenswrapper[4857]: E1201 21:36:30.834661 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:31 crc kubenswrapper[4857]: I1201 21:36:31.834756 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:31 crc kubenswrapper[4857]: I1201 21:36:31.837143 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:31 crc kubenswrapper[4857]: E1201 21:36:31.837252 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:31 crc kubenswrapper[4857]: E1201 21:36:31.837291 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:31 crc kubenswrapper[4857]: I1201 21:36:31.880665 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=7.880648711 podStartE2EDuration="7.880648711s" podCreationTimestamp="2025-12-01 21:36:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:31.878977688 +0000 UTC m=+110.369040005" watchObservedRunningTime="2025-12-01 21:36:31.880648711 +0000 UTC m=+110.370711028" Dec 01 21:36:32 crc kubenswrapper[4857]: I1201 21:36:32.834665 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:32 crc kubenswrapper[4857]: I1201 21:36:32.834684 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:32 crc kubenswrapper[4857]: E1201 21:36:32.834789 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:32 crc kubenswrapper[4857]: E1201 21:36:32.834894 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:33 crc kubenswrapper[4857]: I1201 21:36:33.834265 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:33 crc kubenswrapper[4857]: I1201 21:36:33.834317 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:33 crc kubenswrapper[4857]: E1201 21:36:33.835413 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:33 crc kubenswrapper[4857]: E1201 21:36:33.835302 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:33 crc kubenswrapper[4857]: I1201 21:36:33.836408 4857 scope.go:117] "RemoveContainer" containerID="9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411" Dec 01 21:36:33 crc kubenswrapper[4857]: E1201 21:36:33.836706 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mdc5b_openshift-ovn-kubernetes(557fd3cd-9ab9-4135-af14-f748fdc34a90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" Dec 01 21:36:34 crc kubenswrapper[4857]: I1201 21:36:34.834213 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:34 crc kubenswrapper[4857]: I1201 21:36:34.834247 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:34 crc kubenswrapper[4857]: E1201 21:36:34.834388 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:34 crc kubenswrapper[4857]: E1201 21:36:34.834513 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:35 crc kubenswrapper[4857]: I1201 21:36:35.834291 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:35 crc kubenswrapper[4857]: E1201 21:36:35.834667 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:35 crc kubenswrapper[4857]: I1201 21:36:35.834465 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:35 crc kubenswrapper[4857]: E1201 21:36:35.834908 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:36 crc kubenswrapper[4857]: I1201 21:36:36.834244 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:36 crc kubenswrapper[4857]: I1201 21:36:36.834268 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:36 crc kubenswrapper[4857]: E1201 21:36:36.834430 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:36 crc kubenswrapper[4857]: E1201 21:36:36.834694 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:37 crc kubenswrapper[4857]: I1201 21:36:37.834262 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:37 crc kubenswrapper[4857]: I1201 21:36:37.834342 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:37 crc kubenswrapper[4857]: E1201 21:36:37.834422 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:37 crc kubenswrapper[4857]: E1201 21:36:37.834953 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:38 crc kubenswrapper[4857]: I1201 21:36:38.834092 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:38 crc kubenswrapper[4857]: I1201 21:36:38.834141 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:38 crc kubenswrapper[4857]: E1201 21:36:38.834292 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:38 crc kubenswrapper[4857]: E1201 21:36:38.834449 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:39 crc kubenswrapper[4857]: I1201 21:36:39.834368 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:39 crc kubenswrapper[4857]: I1201 21:36:39.834369 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:39 crc kubenswrapper[4857]: E1201 21:36:39.834627 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:39 crc kubenswrapper[4857]: E1201 21:36:39.834508 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:40 crc kubenswrapper[4857]: I1201 21:36:40.534460 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dfxtl_fc7d55e5-8085-40b7-8147-8508ee6c3f61/kube-multus/1.log" Dec 01 21:36:40 crc kubenswrapper[4857]: I1201 21:36:40.535110 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dfxtl_fc7d55e5-8085-40b7-8147-8508ee6c3f61/kube-multus/0.log" Dec 01 21:36:40 crc kubenswrapper[4857]: I1201 21:36:40.535163 4857 generic.go:334] "Generic (PLEG): container finished" podID="fc7d55e5-8085-40b7-8147-8508ee6c3f61" containerID="6e38c5755141156082b254a121520d58f44c5cef43ad63c3cadb2f1cf8e186e7" exitCode=1 Dec 01 21:36:40 crc kubenswrapper[4857]: I1201 21:36:40.535207 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dfxtl" event={"ID":"fc7d55e5-8085-40b7-8147-8508ee6c3f61","Type":"ContainerDied","Data":"6e38c5755141156082b254a121520d58f44c5cef43ad63c3cadb2f1cf8e186e7"} Dec 01 21:36:40 crc kubenswrapper[4857]: I1201 21:36:40.535248 4857 scope.go:117] "RemoveContainer" containerID="8e066fb8641b349457a7efcf3d1fbb8dbe147d929e2098596e1a2e01be4880d3" Dec 01 21:36:40 crc kubenswrapper[4857]: I1201 21:36:40.536174 4857 scope.go:117] "RemoveContainer" containerID="6e38c5755141156082b254a121520d58f44c5cef43ad63c3cadb2f1cf8e186e7" Dec 01 21:36:40 crc kubenswrapper[4857]: E1201 21:36:40.536587 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-dfxtl_openshift-multus(fc7d55e5-8085-40b7-8147-8508ee6c3f61)\"" pod="openshift-multus/multus-dfxtl" podUID="fc7d55e5-8085-40b7-8147-8508ee6c3f61" Dec 01 21:36:40 crc kubenswrapper[4857]: I1201 21:36:40.834220 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:40 crc kubenswrapper[4857]: E1201 21:36:40.834740 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:40 crc kubenswrapper[4857]: I1201 21:36:40.834258 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:40 crc kubenswrapper[4857]: E1201 21:36:40.835390 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:41 crc kubenswrapper[4857]: I1201 21:36:41.541925 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dfxtl_fc7d55e5-8085-40b7-8147-8508ee6c3f61/kube-multus/1.log" Dec 01 21:36:41 crc kubenswrapper[4857]: E1201 21:36:41.800670 4857 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 01 21:36:41 crc kubenswrapper[4857]: I1201 21:36:41.834232 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:41 crc kubenswrapper[4857]: I1201 21:36:41.834242 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:41 crc kubenswrapper[4857]: E1201 21:36:41.836199 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:41 crc kubenswrapper[4857]: E1201 21:36:41.836524 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:41 crc kubenswrapper[4857]: E1201 21:36:41.966318 4857 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 21:36:42 crc kubenswrapper[4857]: I1201 21:36:42.834934 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:42 crc kubenswrapper[4857]: I1201 21:36:42.835875 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:42 crc kubenswrapper[4857]: E1201 21:36:42.836066 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:42 crc kubenswrapper[4857]: E1201 21:36:42.837187 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:43 crc kubenswrapper[4857]: I1201 21:36:43.834686 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:43 crc kubenswrapper[4857]: E1201 21:36:43.834851 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:43 crc kubenswrapper[4857]: I1201 21:36:43.835206 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:43 crc kubenswrapper[4857]: E1201 21:36:43.835323 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:44 crc kubenswrapper[4857]: I1201 21:36:44.833858 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:44 crc kubenswrapper[4857]: I1201 21:36:44.833915 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:44 crc kubenswrapper[4857]: E1201 21:36:44.834099 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:44 crc kubenswrapper[4857]: E1201 21:36:44.834268 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:45 crc kubenswrapper[4857]: I1201 21:36:45.834610 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:45 crc kubenswrapper[4857]: E1201 21:36:45.834807 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:45 crc kubenswrapper[4857]: I1201 21:36:45.834912 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:45 crc kubenswrapper[4857]: E1201 21:36:45.835176 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:46 crc kubenswrapper[4857]: I1201 21:36:46.834000 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:46 crc kubenswrapper[4857]: I1201 21:36:46.834080 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:46 crc kubenswrapper[4857]: E1201 21:36:46.834455 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:46 crc kubenswrapper[4857]: E1201 21:36:46.834559 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:46 crc kubenswrapper[4857]: I1201 21:36:46.834790 4857 scope.go:117] "RemoveContainer" containerID="9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411" Dec 01 21:36:46 crc kubenswrapper[4857]: E1201 21:36:46.967580 4857 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 21:36:47 crc kubenswrapper[4857]: I1201 21:36:47.563574 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/3.log" Dec 01 21:36:47 crc kubenswrapper[4857]: I1201 21:36:47.566555 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerStarted","Data":"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca"} Dec 01 21:36:47 crc kubenswrapper[4857]: I1201 21:36:47.567069 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:36:47 crc kubenswrapper[4857]: I1201 21:36:47.604895 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podStartSLOduration=106.604874394 podStartE2EDuration="1m46.604874394s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:36:47.604250718 +0000 UTC m=+126.094313045" watchObservedRunningTime="2025-12-01 21:36:47.604874394 +0000 UTC m=+126.094936751" Dec 01 21:36:47 crc kubenswrapper[4857]: I1201 21:36:47.833970 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:47 crc kubenswrapper[4857]: I1201 21:36:47.834073 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:47 crc kubenswrapper[4857]: E1201 21:36:47.834184 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:47 crc kubenswrapper[4857]: E1201 21:36:47.834407 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:47 crc kubenswrapper[4857]: I1201 21:36:47.923365 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-n2fsz"] Dec 01 21:36:48 crc kubenswrapper[4857]: I1201 21:36:48.570705 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:48 crc kubenswrapper[4857]: E1201 21:36:48.570956 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:48 crc kubenswrapper[4857]: I1201 21:36:48.834836 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:48 crc kubenswrapper[4857]: I1201 21:36:48.834875 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:48 crc kubenswrapper[4857]: E1201 21:36:48.835027 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:48 crc kubenswrapper[4857]: E1201 21:36:48.835222 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:49 crc kubenswrapper[4857]: I1201 21:36:49.834325 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:49 crc kubenswrapper[4857]: E1201 21:36:49.834528 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:49 crc kubenswrapper[4857]: I1201 21:36:49.834958 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:49 crc kubenswrapper[4857]: E1201 21:36:49.835140 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:50 crc kubenswrapper[4857]: I1201 21:36:50.834534 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:50 crc kubenswrapper[4857]: I1201 21:36:50.834555 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:50 crc kubenswrapper[4857]: E1201 21:36:50.834748 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:50 crc kubenswrapper[4857]: E1201 21:36:50.834800 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:51 crc kubenswrapper[4857]: I1201 21:36:51.834841 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:51 crc kubenswrapper[4857]: I1201 21:36:51.835036 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:51 crc kubenswrapper[4857]: E1201 21:36:51.836069 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:51 crc kubenswrapper[4857]: E1201 21:36:51.836169 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:51 crc kubenswrapper[4857]: E1201 21:36:51.970402 4857 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 21:36:52 crc kubenswrapper[4857]: I1201 21:36:52.834256 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:52 crc kubenswrapper[4857]: I1201 21:36:52.834355 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:52 crc kubenswrapper[4857]: E1201 21:36:52.834387 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:52 crc kubenswrapper[4857]: E1201 21:36:52.834540 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:53 crc kubenswrapper[4857]: I1201 21:36:53.834154 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:53 crc kubenswrapper[4857]: E1201 21:36:53.834356 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:53 crc kubenswrapper[4857]: I1201 21:36:53.834820 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:53 crc kubenswrapper[4857]: E1201 21:36:53.834927 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:54 crc kubenswrapper[4857]: I1201 21:36:54.834640 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:54 crc kubenswrapper[4857]: I1201 21:36:54.834700 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:54 crc kubenswrapper[4857]: E1201 21:36:54.835062 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:54 crc kubenswrapper[4857]: E1201 21:36:54.835179 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:54 crc kubenswrapper[4857]: I1201 21:36:54.835353 4857 scope.go:117] "RemoveContainer" containerID="6e38c5755141156082b254a121520d58f44c5cef43ad63c3cadb2f1cf8e186e7" Dec 01 21:36:55 crc kubenswrapper[4857]: I1201 21:36:55.598657 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dfxtl_fc7d55e5-8085-40b7-8147-8508ee6c3f61/kube-multus/1.log" Dec 01 21:36:55 crc kubenswrapper[4857]: I1201 21:36:55.599120 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dfxtl" event={"ID":"fc7d55e5-8085-40b7-8147-8508ee6c3f61","Type":"ContainerStarted","Data":"0d75201bf392092adaf26780aa1582bd7f68dae7e1f0c99d67785a27dd80b8f8"} Dec 01 21:36:55 crc kubenswrapper[4857]: I1201 21:36:55.834798 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:55 crc kubenswrapper[4857]: I1201 21:36:55.834870 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:55 crc kubenswrapper[4857]: E1201 21:36:55.835208 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:55 crc kubenswrapper[4857]: E1201 21:36:55.835486 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:56 crc kubenswrapper[4857]: I1201 21:36:56.834013 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:56 crc kubenswrapper[4857]: I1201 21:36:56.834104 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:56 crc kubenswrapper[4857]: E1201 21:36:56.834248 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:56 crc kubenswrapper[4857]: E1201 21:36:56.834432 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:56 crc kubenswrapper[4857]: E1201 21:36:56.972337 4857 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 21:36:57 crc kubenswrapper[4857]: I1201 21:36:57.835091 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:57 crc kubenswrapper[4857]: I1201 21:36:57.835172 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:57 crc kubenswrapper[4857]: E1201 21:36:57.835306 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:57 crc kubenswrapper[4857]: E1201 21:36:57.835414 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:36:58 crc kubenswrapper[4857]: I1201 21:36:58.834210 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:36:58 crc kubenswrapper[4857]: I1201 21:36:58.834211 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:36:58 crc kubenswrapper[4857]: E1201 21:36:58.834550 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:36:58 crc kubenswrapper[4857]: E1201 21:36:58.834415 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:36:59 crc kubenswrapper[4857]: I1201 21:36:59.834938 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:36:59 crc kubenswrapper[4857]: I1201 21:36:59.834972 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:36:59 crc kubenswrapper[4857]: E1201 21:36:59.835218 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:36:59 crc kubenswrapper[4857]: E1201 21:36:59.835343 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:37:00 crc kubenswrapper[4857]: I1201 21:37:00.834030 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:37:00 crc kubenswrapper[4857]: E1201 21:37:00.834215 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 21:37:00 crc kubenswrapper[4857]: I1201 21:37:00.834258 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:37:00 crc kubenswrapper[4857]: E1201 21:37:00.834410 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 21:37:01 crc kubenswrapper[4857]: I1201 21:37:01.835390 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:37:01 crc kubenswrapper[4857]: I1201 21:37:01.835541 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:37:01 crc kubenswrapper[4857]: E1201 21:37:01.836491 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n2fsz" podUID="485d0d95-8002-4bc3-8935-8b8f4ab5b2b3" Dec 01 21:37:01 crc kubenswrapper[4857]: E1201 21:37:01.836621 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 21:37:02 crc kubenswrapper[4857]: I1201 21:37:02.834407 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:37:02 crc kubenswrapper[4857]: I1201 21:37:02.834424 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:37:02 crc kubenswrapper[4857]: I1201 21:37:02.837263 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 21:37:02 crc kubenswrapper[4857]: I1201 21:37:02.841015 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 21:37:03 crc kubenswrapper[4857]: I1201 21:37:03.834783 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:37:03 crc kubenswrapper[4857]: I1201 21:37:03.834932 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:37:03 crc kubenswrapper[4857]: I1201 21:37:03.838440 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 21:37:03 crc kubenswrapper[4857]: I1201 21:37:03.839084 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 21:37:03 crc kubenswrapper[4857]: I1201 21:37:03.839099 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 21:37:03 crc kubenswrapper[4857]: I1201 21:37:03.839411 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.152647 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:08 crc kubenswrapper[4857]: E1201 21:37:08.152855 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:39:10.152807557 +0000 UTC m=+268.642869914 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.167198 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.222275 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7x54g"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.223469 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pcsb9"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.223483 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.224345 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.230812 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.231741 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.231904 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-r5kwb"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.232992 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.233460 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.233699 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.236822 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.237646 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.238826 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.241980 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.242337 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.272784 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291059 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5sxn\" (UniqueName: \"kubernetes.io/projected/5c2a891e-2e39-4fc8-b694-129e008df4b7-kube-api-access-p5sxn\") pod \"cluster-samples-operator-665b6dd947-mmv46\" (UID: \"5c2a891e-2e39-4fc8-b694-129e008df4b7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291113 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-oauth-serving-cert\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291149 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27hc5\" (UniqueName: \"kubernetes.io/projected/b8e5475f-0369-448a-895e-c02b78884677-kube-api-access-27hc5\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291170 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct4v8\" (UniqueName: \"kubernetes.io/projected/d7023a50-1a22-4406-8080-d373309e2155-kube-api-access-ct4v8\") pod \"cluster-image-registry-operator-dc59b4c8b-rxh8n\" (UID: \"d7023a50-1a22-4406-8080-d373309e2155\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291190 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5c2a891e-2e39-4fc8-b694-129e008df4b7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-mmv46\" (UID: \"5c2a891e-2e39-4fc8-b694-129e008df4b7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291214 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291232 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/468167af-db29-49a3-ac86-d4e950d01293-serving-cert\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291254 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d7023a50-1a22-4406-8080-d373309e2155-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rxh8n\" (UID: \"d7023a50-1a22-4406-8080-d373309e2155\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291274 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-console-config\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291294 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291317 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5952edee-93a8-4b8c-9fa0-e15699b86f5d-serving-cert\") pod \"openshift-config-operator-7777fb866f-7x54g\" (UID: \"5952edee-93a8-4b8c-9fa0-e15699b86f5d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291336 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/468167af-db29-49a3-ac86-d4e950d01293-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291358 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d7023a50-1a22-4406-8080-d373309e2155-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rxh8n\" (UID: \"d7023a50-1a22-4406-8080-d373309e2155\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291374 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b8e5475f-0369-448a-895e-c02b78884677-console-oauth-config\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291422 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdxd9\" (UniqueName: \"kubernetes.io/projected/468167af-db29-49a3-ac86-d4e950d01293-kube-api-access-cdxd9\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291447 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291466 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5952edee-93a8-4b8c-9fa0-e15699b86f5d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7x54g\" (UID: \"5952edee-93a8-4b8c-9fa0-e15699b86f5d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291487 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/468167af-db29-49a3-ac86-d4e950d01293-config\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291509 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-trusted-ca-bundle\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291529 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-service-ca\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291551 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/468167af-db29-49a3-ac86-d4e950d01293-service-ca-bundle\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291570 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291598 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwq4c\" (UniqueName: \"kubernetes.io/projected/5952edee-93a8-4b8c-9fa0-e15699b86f5d-kube-api-access-mwq4c\") pod \"openshift-config-operator-7777fb866f-7x54g\" (UID: \"5952edee-93a8-4b8c-9fa0-e15699b86f5d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291625 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d7023a50-1a22-4406-8080-d373309e2155-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rxh8n\" (UID: \"d7023a50-1a22-4406-8080-d373309e2155\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.291646 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b8e5475f-0369-448a-895e-c02b78884677-console-serving-cert\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.297189 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.297857 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.298071 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.298191 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.298301 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.298410 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.298898 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.299073 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.300226 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.300259 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.300437 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.300567 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.300685 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.300795 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.301026 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.301377 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.301640 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.301875 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.301988 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.302934 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.303021 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.308500 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.308524 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.308687 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.309480 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.324447 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.324900 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.324952 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-g6p59"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.325857 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.328227 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2mhm7"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.328592 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mmb9s"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.328859 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-w54ht"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.329204 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-w54ht" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.329213 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.329457 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.329590 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.329695 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.329728 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.329850 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.329974 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.330117 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5wggv"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.330568 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-2bwxl"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.330980 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.331304 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.331359 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.332060 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.334538 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.335255 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.352962 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.353253 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.353485 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.353613 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.354007 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.353631 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.355143 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.353672 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.356759 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.353732 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.357199 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pcsb9"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.351498 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.353775 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.361388 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.362783 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.362901 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.363401 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.365977 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zdzg5"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.368265 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.369492 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7x54g"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.369511 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.369721 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.369888 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.370075 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.370166 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.370433 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.370662 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.370681 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.370807 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.371018 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.371108 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.371400 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.371564 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.371637 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.372028 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.372248 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.373615 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.374614 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.374907 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.374936 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.375076 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.375142 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.375335 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.375464 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.375514 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.375656 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.375723 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.375831 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.375855 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.379433 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.379677 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.379783 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.380058 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.380245 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.380099 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.380509 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7g4cw"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.380606 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.380865 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.381395 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.381495 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.382271 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.382526 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.382590 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.382748 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.383079 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.383195 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.383298 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.383444 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.383540 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4mvhq"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.383558 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.384327 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rg5g2"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.384985 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rg5g2" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.385244 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7g4cw" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.385438 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.389879 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.389960 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.389907 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.390147 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.390387 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.390549 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.390676 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.406414 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.408012 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6875l"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.408953 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6875l" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.409426 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.410364 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.410482 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412017 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412240 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5952edee-93a8-4b8c-9fa0-e15699b86f5d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7x54g\" (UID: \"5952edee-93a8-4b8c-9fa0-e15699b86f5d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412281 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412287 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412303 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/468167af-db29-49a3-ac86-d4e950d01293-config\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412504 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-service-ca\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412539 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-trusted-ca-bundle\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412575 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71476481-b2d8-4a26-b8a8-66fa5ec7135e-serving-cert\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412600 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/468167af-db29-49a3-ac86-d4e950d01293-service-ca-bundle\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412618 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d26df1fc-905c-4a0e-8933-fabccc185cbc-config\") pod \"console-operator-58897d9998-2mhm7\" (UID: \"d26df1fc-905c-4a0e-8933-fabccc185cbc\") " pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412642 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/334d54e0-4413-43e2-875d-bbca874149be-encryption-config\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412667 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ff6afde-262c-4ebf-be69-2e043814271c-audit-dir\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412688 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwq4c\" (UniqueName: \"kubernetes.io/projected/5952edee-93a8-4b8c-9fa0-e15699b86f5d-kube-api-access-mwq4c\") pod \"openshift-config-operator-7777fb866f-7x54g\" (UID: \"5952edee-93a8-4b8c-9fa0-e15699b86f5d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412706 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/334d54e0-4413-43e2-875d-bbca874149be-audit-policies\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412725 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/00d4b52b-5372-48a1-a1b4-6c17e18cc7a4-images\") pod \"machine-api-operator-5694c8668f-2bwxl\" (UID: \"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412752 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412773 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d26df1fc-905c-4a0e-8933-fabccc185cbc-serving-cert\") pod \"console-operator-58897d9998-2mhm7\" (UID: \"d26df1fc-905c-4a0e-8933-fabccc185cbc\") " pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412764 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5952edee-93a8-4b8c-9fa0-e15699b86f5d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7x54g\" (UID: \"5952edee-93a8-4b8c-9fa0-e15699b86f5d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412799 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d7023a50-1a22-4406-8080-d373309e2155-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rxh8n\" (UID: \"d7023a50-1a22-4406-8080-d373309e2155\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412899 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b8e5475f-0369-448a-895e-c02b78884677-console-serving-cert\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412958 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d4b52b-5372-48a1-a1b4-6c17e18cc7a4-config\") pod \"machine-api-operator-5694c8668f-2bwxl\" (UID: \"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412967 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/468167af-db29-49a3-ac86-d4e950d01293-config\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.412978 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/369ee7eb-c09c-4406-90de-5c3a4442d0c5-config\") pod \"machine-approver-56656f9798-fs9lv\" (UID: \"369ee7eb-c09c-4406-90de-5c3a4442d0c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.413005 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-client-ca\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.413739 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/468167af-db29-49a3-ac86-d4e950d01293-service-ca-bundle\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.413957 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-service-ca\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.413971 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p8gq\" (UniqueName: \"kubernetes.io/projected/17e47df6-fb87-44b9-a1a8-143d629f900a-kube-api-access-6p8gq\") pod \"dns-operator-744455d44c-7g4cw\" (UID: \"17e47df6-fb87-44b9-a1a8-143d629f900a\") " pod="openshift-dns-operator/dns-operator-744455d44c-7g4cw" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.414031 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsknc\" (UniqueName: \"kubernetes.io/projected/7ff6afde-262c-4ebf-be69-2e043814271c-kube-api-access-hsknc\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.414143 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5sxn\" (UniqueName: \"kubernetes.io/projected/5c2a891e-2e39-4fc8-b694-129e008df4b7-kube-api-access-p5sxn\") pod \"cluster-samples-operator-665b6dd947-mmv46\" (UID: \"5c2a891e-2e39-4fc8-b694-129e008df4b7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.414190 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-oauth-serving-cert\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.414215 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/334d54e0-4413-43e2-875d-bbca874149be-etcd-client\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.414258 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/334d54e0-4413-43e2-875d-bbca874149be-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.414434 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d7023a50-1a22-4406-8080-d373309e2155-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rxh8n\" (UID: \"d7023a50-1a22-4406-8080-d373309e2155\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.414468 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.414655 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.415260 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-oauth-serving-cert\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.419733 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.419736 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.421932 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-trusted-ca-bundle\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422151 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4hfp\" (UniqueName: \"kubernetes.io/projected/d26df1fc-905c-4a0e-8933-fabccc185cbc-kube-api-access-b4hfp\") pod \"console-operator-58897d9998-2mhm7\" (UID: \"d26df1fc-905c-4a0e-8933-fabccc185cbc\") " pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422265 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9jlm\" (UniqueName: \"kubernetes.io/projected/1927b12d-f48b-4ac4-b61a-ca2cc610312b-kube-api-access-h9jlm\") pod \"openshift-controller-manager-operator-756b6f6bc6-wgv2p\" (UID: \"1927b12d-f48b-4ac4-b61a-ca2cc610312b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422319 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/334d54e0-4413-43e2-875d-bbca874149be-audit-dir\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422343 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c55767b-b626-4e74-a0fe-23d71cf0e347-client-ca\") pod \"route-controller-manager-6576b87f9c-8rjjr\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422402 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422429 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59bvx\" (UniqueName: \"kubernetes.io/projected/00d4b52b-5372-48a1-a1b4-6c17e18cc7a4-kube-api-access-59bvx\") pod \"machine-api-operator-5694c8668f-2bwxl\" (UID: \"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422464 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/334d54e0-4413-43e2-875d-bbca874149be-serving-cert\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422490 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlljj\" (UniqueName: \"kubernetes.io/projected/71476481-b2d8-4a26-b8a8-66fa5ec7135e-kube-api-access-jlljj\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422512 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422538 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422589 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27hc5\" (UniqueName: \"kubernetes.io/projected/b8e5475f-0369-448a-895e-c02b78884677-kube-api-access-27hc5\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422621 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c55767b-b626-4e74-a0fe-23d71cf0e347-config\") pod \"route-controller-manager-6576b87f9c-8rjjr\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422646 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17e47df6-fb87-44b9-a1a8-143d629f900a-metrics-tls\") pod \"dns-operator-744455d44c-7g4cw\" (UID: \"17e47df6-fb87-44b9-a1a8-143d629f900a\") " pod="openshift-dns-operator/dns-operator-744455d44c-7g4cw" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422668 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422699 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d26df1fc-905c-4a0e-8933-fabccc185cbc-trusted-ca\") pod \"console-operator-58897d9998-2mhm7\" (UID: \"d26df1fc-905c-4a0e-8933-fabccc185cbc\") " pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422725 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/369ee7eb-c09c-4406-90de-5c3a4442d0c5-auth-proxy-config\") pod \"machine-approver-56656f9798-fs9lv\" (UID: \"369ee7eb-c09c-4406-90de-5c3a4442d0c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422762 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5c2a891e-2e39-4fc8-b694-129e008df4b7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-mmv46\" (UID: \"5c2a891e-2e39-4fc8-b694-129e008df4b7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422791 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct4v8\" (UniqueName: \"kubernetes.io/projected/d7023a50-1a22-4406-8080-d373309e2155-kube-api-access-ct4v8\") pod \"cluster-image-registry-operator-dc59b4c8b-rxh8n\" (UID: \"d7023a50-1a22-4406-8080-d373309e2155\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422813 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/334d54e0-4413-43e2-875d-bbca874149be-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422841 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422864 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422889 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r5bc\" (UniqueName: \"kubernetes.io/projected/334d54e0-4413-43e2-875d-bbca874149be-kube-api-access-9r5bc\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422911 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c55767b-b626-4e74-a0fe-23d71cf0e347-serving-cert\") pod \"route-controller-manager-6576b87f9c-8rjjr\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422946 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/468167af-db29-49a3-ac86-d4e950d01293-serving-cert\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.422973 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-audit-policies\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423014 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d7023a50-1a22-4406-8080-d373309e2155-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rxh8n\" (UID: \"d7023a50-1a22-4406-8080-d373309e2155\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423036 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-console-config\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423078 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-config\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423135 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m86dd\" (UniqueName: \"kubernetes.io/projected/5c55767b-b626-4e74-a0fe-23d71cf0e347-kube-api-access-m86dd\") pod \"route-controller-manager-6576b87f9c-8rjjr\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423159 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423190 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/369ee7eb-c09c-4406-90de-5c3a4442d0c5-machine-approver-tls\") pod \"machine-approver-56656f9798-fs9lv\" (UID: \"369ee7eb-c09c-4406-90de-5c3a4442d0c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423216 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1927b12d-f48b-4ac4-b61a-ca2cc610312b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wgv2p\" (UID: \"1927b12d-f48b-4ac4-b61a-ca2cc610312b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423257 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5952edee-93a8-4b8c-9fa0-e15699b86f5d-serving-cert\") pod \"openshift-config-operator-7777fb866f-7x54g\" (UID: \"5952edee-93a8-4b8c-9fa0-e15699b86f5d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423287 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423308 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjbvp\" (UniqueName: \"kubernetes.io/projected/369ee7eb-c09c-4406-90de-5c3a4442d0c5-kube-api-access-fjbvp\") pod \"machine-approver-56656f9798-fs9lv\" (UID: \"369ee7eb-c09c-4406-90de-5c3a4442d0c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423330 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1927b12d-f48b-4ac4-b61a-ca2cc610312b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wgv2p\" (UID: \"1927b12d-f48b-4ac4-b61a-ca2cc610312b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423361 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d7023a50-1a22-4406-8080-d373309e2155-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rxh8n\" (UID: \"d7023a50-1a22-4406-8080-d373309e2155\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423382 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/468167af-db29-49a3-ac86-d4e950d01293-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423405 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/00d4b52b-5372-48a1-a1b4-6c17e18cc7a4-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-2bwxl\" (UID: \"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423429 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423460 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b8e5475f-0369-448a-895e-c02b78884677-console-oauth-config\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.423487 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdxd9\" (UniqueName: \"kubernetes.io/projected/468167af-db29-49a3-ac86-d4e950d01293-kube-api-access-cdxd9\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.424510 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-console-config\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.426244 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/468167af-db29-49a3-ac86-d4e950d01293-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.426391 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.427431 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.429339 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/468167af-db29-49a3-ac86-d4e950d01293-serving-cert\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.429522 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.429703 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b8e5475f-0369-448a-895e-c02b78884677-console-serving-cert\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.430244 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-z26dc"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.430823 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.430849 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.431940 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.431918 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5c2a891e-2e39-4fc8-b694-129e008df4b7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-mmv46\" (UID: \"5c2a891e-2e39-4fc8-b694-129e008df4b7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.433195 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.433809 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d7023a50-1a22-4406-8080-d373309e2155-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rxh8n\" (UID: \"d7023a50-1a22-4406-8080-d373309e2155\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.433899 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5952edee-93a8-4b8c-9fa0-e15699b86f5d-serving-cert\") pod \"openshift-config-operator-7777fb866f-7x54g\" (UID: \"5952edee-93a8-4b8c-9fa0-e15699b86f5d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.434368 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.439092 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b8e5475f-0369-448a-895e-c02b78884677-console-oauth-config\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.439407 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.442419 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.443028 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.447192 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.448217 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.448852 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.448982 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.449861 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.451554 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.451741 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.452430 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-r5kwb"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.452442 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.453822 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.454815 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nrp5f"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.455056 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.455388 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.455863 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.456644 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.457114 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.457579 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.465602 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.466109 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.467438 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.469397 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.470155 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.471802 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2mhm7"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.473506 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.480739 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mbzs2"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.481411 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.482077 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.482653 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.482676 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.485613 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mmb9s"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.492049 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.496119 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-2bwxl"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.497410 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-p44n8"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.499156 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-p44n8" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.501905 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.503502 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5wggv"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.504591 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.505692 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.505970 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.506670 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6mll7"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.508509 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-tgwpg"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.509138 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-g6p59"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.509253 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.509591 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.509796 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.512863 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.516345 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7g4cw"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.516436 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.516451 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.519121 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rg5g2"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.519659 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.521815 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-w54ht"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.524195 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6875l"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.524266 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.525642 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.525693 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.525754 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.525789 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f264c9f-c454-4521-a501-6e230287a919-trusted-ca\") pod \"ingress-operator-5b745b69d9-dqzb8\" (UID: \"6f264c9f-c454-4521-a501-6e230287a919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.525836 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.525860 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-serving-cert\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.526071 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pwq6n\" (UID: \"5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.526119 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71476481-b2d8-4a26-b8a8-66fa5ec7135e-serving-cert\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.526192 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-image-import-ca\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.526258 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-audit-dir\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.526287 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/334d54e0-4413-43e2-875d-bbca874149be-encryption-config\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.526337 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ff6afde-262c-4ebf-be69-2e043814271c-audit-dir\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.526364 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d26df1fc-905c-4a0e-8933-fabccc185cbc-config\") pod \"console-operator-58897d9998-2mhm7\" (UID: \"d26df1fc-905c-4a0e-8933-fabccc185cbc\") " pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.526417 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-etcd-serving-ca\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.526444 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6cqq\" (UniqueName: \"kubernetes.io/projected/0fd59b67-77d3-4e4f-b0b4-4bc714e29e73-kube-api-access-d6cqq\") pod \"downloads-7954f5f757-w54ht\" (UID: \"0fd59b67-77d3-4e4f-b0b4-4bc714e29e73\") " pod="openshift-console/downloads-7954f5f757-w54ht" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.526494 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/334d54e0-4413-43e2-875d-bbca874149be-audit-policies\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.526518 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/00d4b52b-5372-48a1-a1b4-6c17e18cc7a4-images\") pod \"machine-api-operator-5694c8668f-2bwxl\" (UID: \"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.526576 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d26df1fc-905c-4a0e-8933-fabccc185cbc-serving-cert\") pod \"console-operator-58897d9998-2mhm7\" (UID: \"d26df1fc-905c-4a0e-8933-fabccc185cbc\") " pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.527260 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.527597 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.527655 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.527767 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pwq6n\" (UID: \"5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.527790 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.527804 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/369ee7eb-c09c-4406-90de-5c3a4442d0c5-config\") pod \"machine-approver-56656f9798-fs9lv\" (UID: \"369ee7eb-c09c-4406-90de-5c3a4442d0c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.527839 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d4b52b-5372-48a1-a1b4-6c17e18cc7a4-config\") pod \"machine-api-operator-5694c8668f-2bwxl\" (UID: \"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.527866 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-client-ca\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.527920 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.527927 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-audit\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.528019 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p8gq\" (UniqueName: \"kubernetes.io/projected/17e47df6-fb87-44b9-a1a8-143d629f900a-kube-api-access-6p8gq\") pod \"dns-operator-744455d44c-7g4cw\" (UID: \"17e47df6-fb87-44b9-a1a8-143d629f900a\") " pod="openshift-dns-operator/dns-operator-744455d44c-7g4cw" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.528061 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsknc\" (UniqueName: \"kubernetes.io/projected/7ff6afde-262c-4ebf-be69-2e043814271c-kube-api-access-hsknc\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.528092 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-config\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.528135 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ff6afde-262c-4ebf-be69-2e043814271c-audit-dir\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.529396 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/334d54e0-4413-43e2-875d-bbca874149be-audit-policies\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.529769 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.528139 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/334d54e0-4413-43e2-875d-bbca874149be-etcd-client\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530237 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/334d54e0-4413-43e2-875d-bbca874149be-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530274 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-etcd-client\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530302 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/334d54e0-4413-43e2-875d-bbca874149be-audit-dir\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530323 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c55767b-b626-4e74-a0fe-23d71cf0e347-client-ca\") pod \"route-controller-manager-6576b87f9c-8rjjr\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530343 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530366 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59bvx\" (UniqueName: \"kubernetes.io/projected/00d4b52b-5372-48a1-a1b4-6c17e18cc7a4-kube-api-access-59bvx\") pod \"machine-api-operator-5694c8668f-2bwxl\" (UID: \"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530384 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4hfp\" (UniqueName: \"kubernetes.io/projected/d26df1fc-905c-4a0e-8933-fabccc185cbc-kube-api-access-b4hfp\") pod \"console-operator-58897d9998-2mhm7\" (UID: \"d26df1fc-905c-4a0e-8933-fabccc185cbc\") " pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530403 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9jlm\" (UniqueName: \"kubernetes.io/projected/1927b12d-f48b-4ac4-b61a-ca2cc610312b-kube-api-access-h9jlm\") pod \"openshift-controller-manager-operator-756b6f6bc6-wgv2p\" (UID: \"1927b12d-f48b-4ac4-b61a-ca2cc610312b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530428 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/334d54e0-4413-43e2-875d-bbca874149be-serving-cert\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530450 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlljj\" (UniqueName: \"kubernetes.io/projected/71476481-b2d8-4a26-b8a8-66fa5ec7135e-kube-api-access-jlljj\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530468 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530490 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530515 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56f48\" (UniqueName: \"kubernetes.io/projected/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-kube-api-access-56f48\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530533 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6f264c9f-c454-4521-a501-6e230287a919-metrics-tls\") pod \"ingress-operator-5b745b69d9-dqzb8\" (UID: \"6f264c9f-c454-4521-a501-6e230287a919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530573 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530590 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d26df1fc-905c-4a0e-8933-fabccc185cbc-trusted-ca\") pod \"console-operator-58897d9998-2mhm7\" (UID: \"d26df1fc-905c-4a0e-8933-fabccc185cbc\") " pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530611 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-node-pullsecrets\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530630 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c55767b-b626-4e74-a0fe-23d71cf0e347-config\") pod \"route-controller-manager-6576b87f9c-8rjjr\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530649 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17e47df6-fb87-44b9-a1a8-143d629f900a-metrics-tls\") pod \"dns-operator-744455d44c-7g4cw\" (UID: \"17e47df6-fb87-44b9-a1a8-143d629f900a\") " pod="openshift-dns-operator/dns-operator-744455d44c-7g4cw" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530676 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/334d54e0-4413-43e2-875d-bbca874149be-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530695 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/369ee7eb-c09c-4406-90de-5c3a4442d0c5-auth-proxy-config\") pod \"machine-approver-56656f9798-fs9lv\" (UID: \"369ee7eb-c09c-4406-90de-5c3a4442d0c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530718 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530716 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-client-ca\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530736 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530800 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r5bc\" (UniqueName: \"kubernetes.io/projected/334d54e0-4413-43e2-875d-bbca874149be-kube-api-access-9r5bc\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530825 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c55767b-b626-4e74-a0fe-23d71cf0e347-serving-cert\") pod \"route-controller-manager-6576b87f9c-8rjjr\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530850 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-audit-policies\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530871 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-config\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530881 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/369ee7eb-c09c-4406-90de-5c3a4442d0c5-config\") pod \"machine-approver-56656f9798-fs9lv\" (UID: \"369ee7eb-c09c-4406-90de-5c3a4442d0c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530888 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m86dd\" (UniqueName: \"kubernetes.io/projected/5c55767b-b626-4e74-a0fe-23d71cf0e347-kube-api-access-m86dd\") pod \"route-controller-manager-6576b87f9c-8rjjr\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530881 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/334d54e0-4413-43e2-875d-bbca874149be-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530913 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530941 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-trusted-ca-bundle\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.530988 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-encryption-config\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531023 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/369ee7eb-c09c-4406-90de-5c3a4442d0c5-machine-approver-tls\") pod \"machine-approver-56656f9798-fs9lv\" (UID: \"369ee7eb-c09c-4406-90de-5c3a4442d0c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531059 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1927b12d-f48b-4ac4-b61a-ca2cc610312b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wgv2p\" (UID: \"1927b12d-f48b-4ac4-b61a-ca2cc610312b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531093 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk9n7\" (UniqueName: \"kubernetes.io/projected/5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842-kube-api-access-nk9n7\") pod \"openshift-apiserver-operator-796bbdcf4f-pwq6n\" (UID: \"5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531146 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531169 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjbvp\" (UniqueName: \"kubernetes.io/projected/369ee7eb-c09c-4406-90de-5c3a4442d0c5-kube-api-access-fjbvp\") pod \"machine-approver-56656f9798-fs9lv\" (UID: \"369ee7eb-c09c-4406-90de-5c3a4442d0c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531218 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1927b12d-f48b-4ac4-b61a-ca2cc610312b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wgv2p\" (UID: \"1927b12d-f48b-4ac4-b61a-ca2cc610312b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531243 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6f264c9f-c454-4521-a501-6e230287a919-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dqzb8\" (UID: \"6f264c9f-c454-4521-a501-6e230287a919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531282 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpdgp\" (UniqueName: \"kubernetes.io/projected/6f264c9f-c454-4521-a501-6e230287a919-kube-api-access-lpdgp\") pod \"ingress-operator-5b745b69d9-dqzb8\" (UID: \"6f264c9f-c454-4521-a501-6e230287a919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531313 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/00d4b52b-5372-48a1-a1b4-6c17e18cc7a4-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-2bwxl\" (UID: \"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531612 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531623 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d26df1fc-905c-4a0e-8933-fabccc185cbc-config\") pod \"console-operator-58897d9998-2mhm7\" (UID: \"d26df1fc-905c-4a0e-8933-fabccc185cbc\") " pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531835 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/334d54e0-4413-43e2-875d-bbca874149be-audit-dir\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.531973 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d26df1fc-905c-4a0e-8933-fabccc185cbc-serving-cert\") pod \"console-operator-58897d9998-2mhm7\" (UID: \"d26df1fc-905c-4a0e-8933-fabccc185cbc\") " pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.532394 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d4b52b-5372-48a1-a1b4-6c17e18cc7a4-config\") pod \"machine-api-operator-5694c8668f-2bwxl\" (UID: \"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.532712 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c55767b-b626-4e74-a0fe-23d71cf0e347-client-ca\") pod \"route-controller-manager-6576b87f9c-8rjjr\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.532852 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.533302 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/00d4b52b-5372-48a1-a1b4-6c17e18cc7a4-images\") pod \"machine-api-operator-5694c8668f-2bwxl\" (UID: \"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.534767 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/334d54e0-4413-43e2-875d-bbca874149be-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.535525 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/369ee7eb-c09c-4406-90de-5c3a4442d0c5-auth-proxy-config\") pod \"machine-approver-56656f9798-fs9lv\" (UID: \"369ee7eb-c09c-4406-90de-5c3a4442d0c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.537335 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c55767b-b626-4e74-a0fe-23d71cf0e347-config\") pod \"route-controller-manager-6576b87f9c-8rjjr\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.538131 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.538189 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.538249 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zdzg5"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.538268 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c55767b-b626-4e74-a0fe-23d71cf0e347-serving-cert\") pod \"route-controller-manager-6576b87f9c-8rjjr\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.538571 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/334d54e0-4413-43e2-875d-bbca874149be-serving-cert\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.538933 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/00d4b52b-5372-48a1-a1b4-6c17e18cc7a4-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-2bwxl\" (UID: \"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.539107 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d26df1fc-905c-4a0e-8933-fabccc185cbc-trusted-ca\") pod \"console-operator-58897d9998-2mhm7\" (UID: \"d26df1fc-905c-4a0e-8933-fabccc185cbc\") " pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.539171 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71476481-b2d8-4a26-b8a8-66fa5ec7135e-serving-cert\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.539673 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.540210 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-config\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.540285 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.540366 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4mvhq"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.540633 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-audit-policies\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.540962 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1927b12d-f48b-4ac4-b61a-ca2cc610312b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wgv2p\" (UID: \"1927b12d-f48b-4ac4-b61a-ca2cc610312b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.541554 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/334d54e0-4413-43e2-875d-bbca874149be-etcd-client\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.541623 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.541949 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1927b12d-f48b-4ac4-b61a-ca2cc610312b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wgv2p\" (UID: \"1927b12d-f48b-4ac4-b61a-ca2cc610312b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.542395 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/369ee7eb-c09c-4406-90de-5c3a4442d0c5-machine-approver-tls\") pod \"machine-approver-56656f9798-fs9lv\" (UID: \"369ee7eb-c09c-4406-90de-5c3a4442d0c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.542706 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.543495 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/334d54e0-4413-43e2-875d-bbca874149be-encryption-config\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.544078 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.544328 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.544378 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.544821 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.545804 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.545847 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.546657 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mbzs2"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.547471 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.548099 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nrp5f"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.548264 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.549448 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.549510 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-qd6km"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.551639 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qd6km"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.551725 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qd6km" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.552029 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-tgwpg"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.552224 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.553430 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6mll7"] Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.561593 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.572814 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.581825 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17e47df6-fb87-44b9-a1a8-143d629f900a-metrics-tls\") pod \"dns-operator-744455d44c-7g4cw\" (UID: \"17e47df6-fb87-44b9-a1a8-143d629f900a\") " pod="openshift-dns-operator/dns-operator-744455d44c-7g4cw" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.586672 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.605686 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.625798 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.636864 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-trusted-ca-bundle\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.636892 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-encryption-config\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.636927 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk9n7\" (UniqueName: \"kubernetes.io/projected/5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842-kube-api-access-nk9n7\") pod \"openshift-apiserver-operator-796bbdcf4f-pwq6n\" (UID: \"5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.636952 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6f264c9f-c454-4521-a501-6e230287a919-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dqzb8\" (UID: \"6f264c9f-c454-4521-a501-6e230287a919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.636970 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpdgp\" (UniqueName: \"kubernetes.io/projected/6f264c9f-c454-4521-a501-6e230287a919-kube-api-access-lpdgp\") pod \"ingress-operator-5b745b69d9-dqzb8\" (UID: \"6f264c9f-c454-4521-a501-6e230287a919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.636991 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f264c9f-c454-4521-a501-6e230287a919-trusted-ca\") pod \"ingress-operator-5b745b69d9-dqzb8\" (UID: \"6f264c9f-c454-4521-a501-6e230287a919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.637361 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-serving-cert\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.637404 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pwq6n\" (UID: \"5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.637428 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-image-import-ca\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.637450 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-audit-dir\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.637467 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-etcd-serving-ca\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.639480 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6cqq\" (UniqueName: \"kubernetes.io/projected/0fd59b67-77d3-4e4f-b0b4-4bc714e29e73-kube-api-access-d6cqq\") pod \"downloads-7954f5f757-w54ht\" (UID: \"0fd59b67-77d3-4e4f-b0b4-4bc714e29e73\") " pod="openshift-console/downloads-7954f5f757-w54ht" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.639648 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pwq6n\" (UID: \"5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.639807 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-audit\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.639960 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-config\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.639997 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-etcd-client\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.640172 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6f264c9f-c454-4521-a501-6e230287a919-metrics-tls\") pod \"ingress-operator-5b745b69d9-dqzb8\" (UID: \"6f264c9f-c454-4521-a501-6e230287a919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.640203 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56f48\" (UniqueName: \"kubernetes.io/projected/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-kube-api-access-56f48\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.640243 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-node-pullsecrets\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.640372 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-node-pullsecrets\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.641435 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-trusted-ca-bundle\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.642339 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-etcd-serving-ca\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.642738 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pwq6n\" (UID: \"5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.642746 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-config\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.642843 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-audit-dir\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.642894 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f264c9f-c454-4521-a501-6e230287a919-trusted-ca\") pod \"ingress-operator-5b745b69d9-dqzb8\" (UID: \"6f264c9f-c454-4521-a501-6e230287a919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.645221 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-audit\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.646433 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pwq6n\" (UID: \"5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.647956 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-image-import-ca\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.649385 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.654405 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6f264c9f-c454-4521-a501-6e230287a919-metrics-tls\") pod \"ingress-operator-5b745b69d9-dqzb8\" (UID: \"6f264c9f-c454-4521-a501-6e230287a919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.659998 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-encryption-config\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.664152 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-serving-cert\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.668974 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-etcd-client\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.670910 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.686053 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.706342 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.727162 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 21:37:08 crc kubenswrapper[4857]: W1201 21:37:08.739088 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-d8bf723fe5c453ab3c59e0526e452e7d404c04f1a21a43d4890f46f213f17f6f WatchSource:0}: Error finding container d8bf723fe5c453ab3c59e0526e452e7d404c04f1a21a43d4890f46f213f17f6f: Status 404 returned error can't find the container with id d8bf723fe5c453ab3c59e0526e452e7d404c04f1a21a43d4890f46f213f17f6f Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.746539 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.770079 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.788186 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.806221 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: W1201 21:37:08.817970 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-12169ce6f1f94f0f8ef19fbcd787458adefd723de77099c84f0f22afbc794eca WatchSource:0}: Error finding container 12169ce6f1f94f0f8ef19fbcd787458adefd723de77099c84f0f22afbc794eca: Status 404 returned error can't find the container with id 12169ce6f1f94f0f8ef19fbcd787458adefd723de77099c84f0f22afbc794eca Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.825797 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.846668 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.909183 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwq4c\" (UniqueName: \"kubernetes.io/projected/5952edee-93a8-4b8c-9fa0-e15699b86f5d-kube-api-access-mwq4c\") pod \"openshift-config-operator-7777fb866f-7x54g\" (UID: \"5952edee-93a8-4b8c-9fa0-e15699b86f5d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.925868 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5sxn\" (UniqueName: \"kubernetes.io/projected/5c2a891e-2e39-4fc8-b694-129e008df4b7-kube-api-access-p5sxn\") pod \"cluster-samples-operator-665b6dd947-mmv46\" (UID: \"5c2a891e-2e39-4fc8-b694-129e008df4b7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.958561 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.971542 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdxd9\" (UniqueName: \"kubernetes.io/projected/468167af-db29-49a3-ac86-d4e950d01293-kube-api-access-cdxd9\") pod \"authentication-operator-69f744f599-pcsb9\" (UID: \"468167af-db29-49a3-ac86-d4e950d01293\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.981900 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27hc5\" (UniqueName: \"kubernetes.io/projected/b8e5475f-0369-448a-895e-c02b78884677-kube-api-access-27hc5\") pod \"console-f9d7485db-r5kwb\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:08 crc kubenswrapper[4857]: I1201 21:37:08.984001 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.005485 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d7023a50-1a22-4406-8080-d373309e2155-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rxh8n\" (UID: \"d7023a50-1a22-4406-8080-d373309e2155\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.022701 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct4v8\" (UniqueName: \"kubernetes.io/projected/d7023a50-1a22-4406-8080-d373309e2155-kube-api-access-ct4v8\") pod \"cluster-image-registry-operator-dc59b4c8b-rxh8n\" (UID: \"d7023a50-1a22-4406-8080-d373309e2155\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.026217 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.048441 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.066898 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.086077 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.107836 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.130317 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.146832 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.165941 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.166633 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.189875 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.199660 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46"] Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.207445 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-r5kwb"] Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.211318 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 21:37:09 crc kubenswrapper[4857]: W1201 21:37:09.220348 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8e5475f_0369_448a_895e_c02b78884677.slice/crio-4dc9776b7b82b01c9f92d944c9c0b1a1da61e7dda60172531293afe70af0ca3e WatchSource:0}: Error finding container 4dc9776b7b82b01c9f92d944c9c0b1a1da61e7dda60172531293afe70af0ca3e: Status 404 returned error can't find the container with id 4dc9776b7b82b01c9f92d944c9c0b1a1da61e7dda60172531293afe70af0ca3e Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.224239 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.226684 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.248122 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.266892 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.286330 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.306884 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.313528 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.326764 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.347281 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.368132 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.386351 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.409230 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.427386 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7x54g"] Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.429521 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.449026 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.468130 4857 request.go:700] Waited for 1.0187881s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmcc-proxy-tls&limit=500&resourceVersion=0 Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.470313 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.492822 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.506794 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.549654 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.550212 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.556538 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pcsb9"] Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.566557 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.586437 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.605760 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.627593 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.647317 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.667497 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.686834 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.692809 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" event={"ID":"468167af-db29-49a3-ac86-d4e950d01293","Type":"ContainerStarted","Data":"59545fb46b6f22bb8fbddad2dfb1d125613d50a186cc7cece919af70de1eb3de"} Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.695191 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" event={"ID":"5952edee-93a8-4b8c-9fa0-e15699b86f5d","Type":"ContainerStarted","Data":"c96e54f451d554dda80eba691278d51a49d2df7f927347469edfebea3264324b"} Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.699418 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46" event={"ID":"5c2a891e-2e39-4fc8-b694-129e008df4b7","Type":"ContainerStarted","Data":"e13c5ed30c0d742d9f96216eb234dfee1d11d967c67a5dbed8c15b037c66085e"} Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.699476 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46" event={"ID":"5c2a891e-2e39-4fc8-b694-129e008df4b7","Type":"ContainerStarted","Data":"09284357d4872692a53ce34b43ea03a14d40fc725a3374912755aa01b9272e0f"} Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.700931 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"fb4fca80454c5e2c4c88619264b1bb26547c397afc3d23097635858320c06ab1"} Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.701006 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d8bf723fe5c453ab3c59e0526e452e7d404c04f1a21a43d4890f46f213f17f6f"} Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.703387 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"11cd2cd1615b4bfc97361437718d12057ee8c49f99fa7c26bb69bc71646dab84"} Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.703438 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ea5ea1dc6a4e5c009061fdb31358d42f56fb4af746a7df9b92b69be8df89d16c"} Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.704298 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.707284 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"390e679ac1a94e74ca67b1dc4edecc8aa41cc86d5c4d38a369f8a6a6f3c6629c"} Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.707312 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"12169ce6f1f94f0f8ef19fbcd787458adefd723de77099c84f0f22afbc794eca"} Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.708908 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r5kwb" event={"ID":"b8e5475f-0369-448a-895e-c02b78884677","Type":"ContainerStarted","Data":"f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f"} Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.708959 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r5kwb" event={"ID":"b8e5475f-0369-448a-895e-c02b78884677","Type":"ContainerStarted","Data":"4dc9776b7b82b01c9f92d944c9c0b1a1da61e7dda60172531293afe70af0ca3e"} Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.732598 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.734698 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.747182 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.766597 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.831448 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.832116 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.832415 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.846674 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.866546 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.950559 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.950899 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.951174 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.952501 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.966912 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 21:37:09 crc kubenswrapper[4857]: I1201 21:37:09.986274 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.005857 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.030273 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.046339 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.067415 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.086818 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.113991 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.126929 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.147427 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.165848 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.171672 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n"] Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.190535 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.214855 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.232509 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.246942 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.266646 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.287859 4857 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.305956 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.349819 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p8gq\" (UniqueName: \"kubernetes.io/projected/17e47df6-fb87-44b9-a1a8-143d629f900a-kube-api-access-6p8gq\") pod \"dns-operator-744455d44c-7g4cw\" (UID: \"17e47df6-fb87-44b9-a1a8-143d629f900a\") " pod="openshift-dns-operator/dns-operator-744455d44c-7g4cw" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.371134 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsknc\" (UniqueName: \"kubernetes.io/projected/7ff6afde-262c-4ebf-be69-2e043814271c-kube-api-access-hsknc\") pod \"oauth-openshift-558db77b4-mmb9s\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.409629 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlljj\" (UniqueName: \"kubernetes.io/projected/71476481-b2d8-4a26-b8a8-66fa5ec7135e-kube-api-access-jlljj\") pod \"controller-manager-879f6c89f-5wggv\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.416094 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59bvx\" (UniqueName: \"kubernetes.io/projected/00d4b52b-5372-48a1-a1b4-6c17e18cc7a4-kube-api-access-59bvx\") pod \"machine-api-operator-5694c8668f-2bwxl\" (UID: \"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.436530 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4hfp\" (UniqueName: \"kubernetes.io/projected/d26df1fc-905c-4a0e-8933-fabccc185cbc-kube-api-access-b4hfp\") pod \"console-operator-58897d9998-2mhm7\" (UID: \"d26df1fc-905c-4a0e-8933-fabccc185cbc\") " pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.455034 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9jlm\" (UniqueName: \"kubernetes.io/projected/1927b12d-f48b-4ac4-b61a-ca2cc610312b-kube-api-access-h9jlm\") pod \"openshift-controller-manager-operator-756b6f6bc6-wgv2p\" (UID: \"1927b12d-f48b-4ac4-b61a-ca2cc610312b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.474535 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7g4cw" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.483613 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m86dd\" (UniqueName: \"kubernetes.io/projected/5c55767b-b626-4e74-a0fe-23d71cf0e347-kube-api-access-m86dd\") pod \"route-controller-manager-6576b87f9c-8rjjr\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.484366 4857 request.go:700] Waited for 1.946015122s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-machine-approver/serviceaccounts/machine-approver-sa/token Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.487212 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r5bc\" (UniqueName: \"kubernetes.io/projected/334d54e0-4413-43e2-875d-bbca874149be-kube-api-access-9r5bc\") pod \"apiserver-7bbb656c7d-wdc4s\" (UID: \"334d54e0-4413-43e2-875d-bbca874149be\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.504964 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjbvp\" (UniqueName: \"kubernetes.io/projected/369ee7eb-c09c-4406-90de-5c3a4442d0c5-kube-api-access-fjbvp\") pod \"machine-approver-56656f9798-fs9lv\" (UID: \"369ee7eb-c09c-4406-90de-5c3a4442d0c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.507008 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.525889 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.549089 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.566119 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.590250 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.600389 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.601329 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6cqq\" (UniqueName: \"kubernetes.io/projected/0fd59b67-77d3-4e4f-b0b4-4bc714e29e73-kube-api-access-d6cqq\") pod \"downloads-7954f5f757-w54ht\" (UID: \"0fd59b67-77d3-4e4f-b0b4-4bc714e29e73\") " pod="openshift-console/downloads-7954f5f757-w54ht" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.605536 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.615437 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.623006 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.624980 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk9n7\" (UniqueName: \"kubernetes.io/projected/5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842-kube-api-access-nk9n7\") pod \"openshift-apiserver-operator-796bbdcf4f-pwq6n\" (UID: \"5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.643757 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.650972 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.657442 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6f264c9f-c454-4521-a501-6e230287a919-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dqzb8\" (UID: \"6f264c9f-c454-4521-a501-6e230287a919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.688016 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpdgp\" (UniqueName: \"kubernetes.io/projected/6f264c9f-c454-4521-a501-6e230287a919-kube-api-access-lpdgp\") pod \"ingress-operator-5b745b69d9-dqzb8\" (UID: \"6f264c9f-c454-4521-a501-6e230287a919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.709508 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.722148 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56f48\" (UniqueName: \"kubernetes.io/projected/b6911089-a8c9-4e07-8f85-5d5d2ff9f736-kube-api-access-56f48\") pod \"apiserver-76f77b778f-g6p59\" (UID: \"b6911089-a8c9-4e07-8f85-5d5d2ff9f736\") " pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.738154 4857 generic.go:334] "Generic (PLEG): container finished" podID="5952edee-93a8-4b8c-9fa0-e15699b86f5d" containerID="f2400143614ae2498f810a8a95074f34b2afb2fb2186c38be27f193ce00df6cf" exitCode=0 Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.738223 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" event={"ID":"5952edee-93a8-4b8c-9fa0-e15699b86f5d","Type":"ContainerDied","Data":"f2400143614ae2498f810a8a95074f34b2afb2fb2186c38be27f193ce00df6cf"} Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.739894 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46" event={"ID":"5c2a891e-2e39-4fc8-b694-129e008df4b7","Type":"ContainerStarted","Data":"cbe1568f9e0818f450373fcaf3e2d0cde23d0f3e23d4d1560eeee2043d1dc32b"} Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.740780 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" event={"ID":"369ee7eb-c09c-4406-90de-5c3a4442d0c5","Type":"ContainerStarted","Data":"5d6410f8c75f36232e7d2ab3245bbd0e89dee5ccdeffe69e09bc93b2a5a64030"} Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.741874 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" event={"ID":"468167af-db29-49a3-ac86-d4e950d01293","Type":"ContainerStarted","Data":"7807287fc26e233d6d91e6d6f2c039ba6732feec2ac8b9e97937d1597cd8f9cb"} Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.743238 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.753916 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" event={"ID":"d7023a50-1a22-4406-8080-d373309e2155","Type":"ContainerStarted","Data":"7baf3d9ae0a4abd70f944afecf612e97cf172f4e47157858d1b3282f6146efa9"} Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.753945 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" event={"ID":"d7023a50-1a22-4406-8080-d373309e2155","Type":"ContainerStarted","Data":"ad25b7a69dcb9df955c630d50d5a35b81e8ba67f8fb06d129d05a474c48778e8"} Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794401 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3301b37b-d60f-42d2-91dd-c0c898019831-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rg5g2\" (UID: \"3301b37b-d60f-42d2-91dd-c0c898019831\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rg5g2" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794436 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-etcd-client\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794467 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-etcd-service-ca\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794504 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794522 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp9pd\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-kube-api-access-zp9pd\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794538 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-config\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794561 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/658992cc-be88-49f4-b35c-bc445583fc39-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794577 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xghgp\" (UniqueName: \"kubernetes.io/projected/3301b37b-d60f-42d2-91dd-c0c898019831-kube-api-access-xghgp\") pod \"multus-admission-controller-857f4d67dd-rg5g2\" (UID: \"3301b37b-d60f-42d2-91dd-c0c898019831\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rg5g2" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794595 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/658992cc-be88-49f4-b35c-bc445583fc39-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794614 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-registry-tls\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794630 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/658992cc-be88-49f4-b35c-bc445583fc39-trusted-ca\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794649 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h69dr\" (UniqueName: \"kubernetes.io/projected/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-kube-api-access-h69dr\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794665 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/658992cc-be88-49f4-b35c-bc445583fc39-registry-certificates\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794682 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-serving-cert\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794699 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-bound-sa-token\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794713 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-etcd-ca\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.794734 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qnkc\" (UniqueName: \"kubernetes.io/projected/e3c0a9de-bb5f-44d5-b99f-f7acd1d0d308-kube-api-access-9qnkc\") pod \"migrator-59844c95c7-6875l\" (UID: \"e3c0a9de-bb5f-44d5-b99f-f7acd1d0d308\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6875l" Dec 01 21:37:10 crc kubenswrapper[4857]: E1201 21:37:10.795486 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:11.29546913 +0000 UTC m=+149.785531467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.875707 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.876480 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.886256 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-w54ht" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.904636 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.904770 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0e090c12-a85e-4a29-9698-98e00f798808-proxy-tls\") pod \"machine-config-operator-74547568cd-jp84x\" (UID: \"0e090c12-a85e-4a29-9698-98e00f798808\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.904792 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csj66\" (UniqueName: \"kubernetes.io/projected/e780f3c5-76bd-47b9-97ff-89a80759f47c-kube-api-access-csj66\") pod \"catalog-operator-68c6474976-kkf64\" (UID: \"e780f3c5-76bd-47b9-97ff-89a80759f47c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.904807 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-plugins-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.904834 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-config\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.904866 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-mountpoint-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.904881 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6590c60-f467-48a5-8551-0599bf6cb370-config-volume\") pod \"dns-default-tgwpg\" (UID: \"d6590c60-f467-48a5-8551-0599bf6cb370\") " pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.904896 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01903422-1e15-44cf-87f1-6037f6631592-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mbzs2\" (UID: \"01903422-1e15-44cf-87f1-6037f6631592\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.904937 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7f9d27b-4546-443e-aecc-d2478cb1a5ae-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-242k4\" (UID: \"e7f9d27b-4546-443e-aecc-d2478cb1a5ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.904950 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e780f3c5-76bd-47b9-97ff-89a80759f47c-srv-cert\") pod \"catalog-operator-68c6474976-kkf64\" (UID: \"e780f3c5-76bd-47b9-97ff-89a80759f47c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.904990 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xghgp\" (UniqueName: \"kubernetes.io/projected/3301b37b-d60f-42d2-91dd-c0c898019831-kube-api-access-xghgp\") pod \"multus-admission-controller-857f4d67dd-rg5g2\" (UID: \"3301b37b-d60f-42d2-91dd-c0c898019831\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rg5g2" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905012 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/658992cc-be88-49f4-b35c-bc445583fc39-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905051 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbpk9\" (UniqueName: \"kubernetes.io/projected/5361e52d-63b2-4f66-b786-9b1348a6b428-kube-api-access-kbpk9\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905070 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c8375eb-1803-475e-b897-9307d7bc29ee-secret-volume\") pod \"collect-profiles-29410410-9gpfb\" (UID: \"6c8375eb-1803-475e-b897-9307d7bc29ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905100 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc175021-8815-4fd7-b42c-af7e29b281fd-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mfv8h\" (UID: \"fc175021-8815-4fd7-b42c-af7e29b281fd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905117 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98b9r\" (UniqueName: \"kubernetes.io/projected/0e090c12-a85e-4a29-9698-98e00f798808-kube-api-access-98b9r\") pod \"machine-config-operator-74547568cd-jp84x\" (UID: \"0e090c12-a85e-4a29-9698-98e00f798808\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905132 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c8375eb-1803-475e-b897-9307d7bc29ee-config-volume\") pod \"collect-profiles-29410410-9gpfb\" (UID: \"6c8375eb-1803-475e-b897-9307d7bc29ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905157 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0e090c12-a85e-4a29-9698-98e00f798808-auth-proxy-config\") pod \"machine-config-operator-74547568cd-jp84x\" (UID: \"0e090c12-a85e-4a29-9698-98e00f798808\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905179 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/658992cc-be88-49f4-b35c-bc445583fc39-trusted-ca\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905202 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc175021-8815-4fd7-b42c-af7e29b281fd-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mfv8h\" (UID: \"fc175021-8815-4fd7-b42c-af7e29b281fd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905225 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h69dr\" (UniqueName: \"kubernetes.io/projected/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-kube-api-access-h69dr\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905276 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d668b26f-a9b6-4544-abf9-64e72728e342-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mxkzn\" (UID: \"d668b26f-a9b6-4544-abf9-64e72728e342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905292 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-serving-cert\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905310 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htc9q\" (UniqueName: \"kubernetes.io/projected/1cb66d76-6c0a-48c9-a0a8-54a60f7aae01-kube-api-access-htc9q\") pod \"control-plane-machine-set-operator-78cbb6b69f-8mmlp\" (UID: \"1cb66d76-6c0a-48c9-a0a8-54a60f7aae01\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905326 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/01903422-1e15-44cf-87f1-6037f6631592-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mbzs2\" (UID: \"01903422-1e15-44cf-87f1-6037f6631592\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905362 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2rr8\" (UniqueName: \"kubernetes.io/projected/1f215aad-96e0-4995-af49-99d7b947e674-kube-api-access-h2rr8\") pod \"ingress-canary-qd6km\" (UID: \"1f215aad-96e0-4995-af49-99d7b947e674\") " pod="openshift-ingress-canary/ingress-canary-qd6km" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905378 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq6hw\" (UniqueName: \"kubernetes.io/projected/6c8375eb-1803-475e-b897-9307d7bc29ee-kube-api-access-nq6hw\") pod \"collect-profiles-29410410-9gpfb\" (UID: \"6c8375eb-1803-475e-b897-9307d7bc29ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905393 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-bound-sa-token\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905408 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt57p\" (UniqueName: \"kubernetes.io/projected/6c4ff7e6-723d-4085-9bbc-bc865c8e6f61-kube-api-access-rt57p\") pod \"packageserver-d55dfcdfc-wg2k4\" (UID: \"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905444 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed50d226-e00b-447f-be13-e5847b1cf84b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6ls6r\" (UID: \"ed50d226-e00b-447f-be13-e5847b1cf84b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905507 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3301b37b-d60f-42d2-91dd-c0c898019831-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rg5g2\" (UID: \"3301b37b-d60f-42d2-91dd-c0c898019831\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rg5g2" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905525 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-etcd-client\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905543 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/555f0013-cb67-45f1-b96b-3f2b938fba11-service-ca-bundle\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905588 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-socket-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.905603 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-csi-data-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:10 crc kubenswrapper[4857]: E1201 21:37:10.907348 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:11.407322742 +0000 UTC m=+149.897385059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.908063 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-config\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.917910 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3301b37b-d60f-42d2-91dd-c0c898019831-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rg5g2\" (UID: \"3301b37b-d60f-42d2-91dd-c0c898019831\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rg5g2" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.918417 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-etcd-client\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.946802 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-etcd-service-ca\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.947947 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc175021-8815-4fd7-b42c-af7e29b281fd-config\") pod \"kube-apiserver-operator-766d6c64bb-mfv8h\" (UID: \"fc175021-8815-4fd7-b42c-af7e29b281fd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.947986 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f215aad-96e0-4995-af49-99d7b947e674-cert\") pod \"ingress-canary-qd6km\" (UID: \"1f215aad-96e0-4995-af49-99d7b947e674\") " pod="openshift-ingress-canary/ingress-canary-qd6km" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948013 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b1d82f9-f291-417b-bf3a-e8085c0a8f2b-serving-cert\") pod \"service-ca-operator-777779d784-jt8d4\" (UID: \"4b1d82f9-f291-417b-bf3a-e8085c0a8f2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948085 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhr5c\" (UniqueName: \"kubernetes.io/projected/e7f9d27b-4546-443e-aecc-d2478cb1a5ae-kube-api-access-vhr5c\") pod \"kube-storage-version-migrator-operator-b67b599dd-242k4\" (UID: \"e7f9d27b-4546-443e-aecc-d2478cb1a5ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948111 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6c4ff7e6-723d-4085-9bbc-bc865c8e6f61-tmpfs\") pod \"packageserver-d55dfcdfc-wg2k4\" (UID: \"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948142 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d668b26f-a9b6-4544-abf9-64e72728e342-proxy-tls\") pod \"machine-config-controller-84d6567774-mxkzn\" (UID: \"d668b26f-a9b6-4544-abf9-64e72728e342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948257 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7f9d27b-4546-443e-aecc-d2478cb1a5ae-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-242k4\" (UID: \"e7f9d27b-4546-443e-aecc-d2478cb1a5ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948284 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1d82f9-f291-417b-bf3a-e8085c0a8f2b-config\") pod \"service-ca-operator-777779d784-jt8d4\" (UID: \"4b1d82f9-f291-417b-bf3a-e8085c0a8f2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948309 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp9pd\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-kube-api-access-zp9pd\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948335 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/1cb66d76-6c0a-48c9-a0a8-54a60f7aae01-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8mmlp\" (UID: \"1cb66d76-6c0a-48c9-a0a8-54a60f7aae01\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948374 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/25436d81-ab74-476f-ae9f-e74ca9452da0-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xrmc2\" (UID: \"25436d81-ab74-476f-ae9f-e74ca9452da0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948395 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6c4ff7e6-723d-4085-9bbc-bc865c8e6f61-apiservice-cert\") pod \"packageserver-d55dfcdfc-wg2k4\" (UID: \"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948432 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4da3afa4-fa1f-42e5-b321-d59419e58910-signing-key\") pod \"service-ca-9c57cc56f-nrp5f\" (UID: \"4da3afa4-fa1f-42e5-b321-d59419e58910\") " pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948503 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4lvj\" (UniqueName: \"kubernetes.io/projected/555f0013-cb67-45f1-b96b-3f2b938fba11-kube-api-access-d4lvj\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948528 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nrtj9\" (UID: \"dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948552 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0e090c12-a85e-4a29-9698-98e00f798808-images\") pod \"machine-config-operator-74547568cd-jp84x\" (UID: \"0e090c12-a85e-4a29-9698-98e00f798808\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948575 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ng9h\" (UniqueName: \"kubernetes.io/projected/01903422-1e15-44cf-87f1-6037f6631592-kube-api-access-5ng9h\") pod \"marketplace-operator-79b997595-mbzs2\" (UID: \"01903422-1e15-44cf-87f1-6037f6631592\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948628 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/658992cc-be88-49f4-b35c-bc445583fc39-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948678 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpjx6\" (UniqueName: \"kubernetes.io/projected/2e0f8071-e17c-4d6f-aefe-771448acbe1a-kube-api-access-mpjx6\") pod \"machine-config-server-p44n8\" (UID: \"2e0f8071-e17c-4d6f-aefe-771448acbe1a\") " pod="openshift-machine-config-operator/machine-config-server-p44n8" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948700 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/25436d81-ab74-476f-ae9f-e74ca9452da0-srv-cert\") pod \"olm-operator-6b444d44fb-xrmc2\" (UID: \"25436d81-ab74-476f-ae9f-e74ca9452da0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948803 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-registry-tls\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948862 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed50d226-e00b-447f-be13-e5847b1cf84b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6ls6r\" (UID: \"ed50d226-e00b-447f-be13-e5847b1cf84b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948884 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmdtc\" (UniqueName: \"kubernetes.io/projected/d668b26f-a9b6-4544-abf9-64e72728e342-kube-api-access-jmdtc\") pod \"machine-config-controller-84d6567774-mxkzn\" (UID: \"d668b26f-a9b6-4544-abf9-64e72728e342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948931 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/555f0013-cb67-45f1-b96b-3f2b938fba11-stats-auth\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.948969 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6590c60-f467-48a5-8551-0599bf6cb370-metrics-tls\") pod \"dns-default-tgwpg\" (UID: \"d6590c60-f467-48a5-8551-0599bf6cb370\") " pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949008 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clc47\" (UniqueName: \"kubernetes.io/projected/4da3afa4-fa1f-42e5-b321-d59419e58910-kube-api-access-clc47\") pod \"service-ca-9c57cc56f-nrp5f\" (UID: \"4da3afa4-fa1f-42e5-b321-d59419e58910\") " pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949083 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/555f0013-cb67-45f1-b96b-3f2b938fba11-default-certificate\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949113 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/658992cc-be88-49f4-b35c-bc445583fc39-registry-certificates\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949148 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nrtj9\" (UID: \"dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949202 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-etcd-ca\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949223 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6c4ff7e6-723d-4085-9bbc-bc865c8e6f61-webhook-cert\") pod \"packageserver-d55dfcdfc-wg2k4\" (UID: \"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949256 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qnkc\" (UniqueName: \"kubernetes.io/projected/e3c0a9de-bb5f-44d5-b99f-f7acd1d0d308-kube-api-access-9qnkc\") pod \"migrator-59844c95c7-6875l\" (UID: \"e3c0a9de-bb5f-44d5-b99f-f7acd1d0d308\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6875l" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949334 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/511408a8-c761-46fe-bb8d-dd1147c0728e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-n5slz\" (UID: \"511408a8-c761-46fe-bb8d-dd1147c0728e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949365 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98czm\" (UniqueName: \"kubernetes.io/projected/25436d81-ab74-476f-ae9f-e74ca9452da0-kube-api-access-98czm\") pod \"olm-operator-6b444d44fb-xrmc2\" (UID: \"25436d81-ab74-476f-ae9f-e74ca9452da0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949388 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9zkf\" (UniqueName: \"kubernetes.io/projected/d6590c60-f467-48a5-8551-0599bf6cb370-kube-api-access-v9zkf\") pod \"dns-default-tgwpg\" (UID: \"d6590c60-f467-48a5-8551-0599bf6cb370\") " pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949469 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-registration-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949494 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mk8s\" (UniqueName: \"kubernetes.io/projected/4b1d82f9-f291-417b-bf3a-e8085c0a8f2b-kube-api-access-9mk8s\") pod \"service-ca-operator-777779d784-jt8d4\" (UID: \"4b1d82f9-f291-417b-bf3a-e8085c0a8f2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949530 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nrtj9\" (UID: \"dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949569 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/2e0f8071-e17c-4d6f-aefe-771448acbe1a-certs\") pod \"machine-config-server-p44n8\" (UID: \"2e0f8071-e17c-4d6f-aefe-771448acbe1a\") " pod="openshift-machine-config-operator/machine-config-server-p44n8" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949634 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/555f0013-cb67-45f1-b96b-3f2b938fba11-metrics-certs\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949657 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/2e0f8071-e17c-4d6f-aefe-771448acbe1a-node-bootstrap-token\") pod \"machine-config-server-p44n8\" (UID: \"2e0f8071-e17c-4d6f-aefe-771448acbe1a\") " pod="openshift-machine-config-operator/machine-config-server-p44n8" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949679 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed50d226-e00b-447f-be13-e5847b1cf84b-config\") pod \"kube-controller-manager-operator-78b949d7b-6ls6r\" (UID: \"ed50d226-e00b-447f-be13-e5847b1cf84b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949816 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xgds\" (UniqueName: \"kubernetes.io/projected/511408a8-c761-46fe-bb8d-dd1147c0728e-kube-api-access-2xgds\") pod \"package-server-manager-789f6589d5-n5slz\" (UID: \"511408a8-c761-46fe-bb8d-dd1147c0728e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949880 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4da3afa4-fa1f-42e5-b321-d59419e58910-signing-cabundle\") pod \"service-ca-9c57cc56f-nrp5f\" (UID: \"4da3afa4-fa1f-42e5-b321-d59419e58910\") " pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.949903 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e780f3c5-76bd-47b9-97ff-89a80759f47c-profile-collector-cert\") pod \"catalog-operator-68c6474976-kkf64\" (UID: \"e780f3c5-76bd-47b9-97ff-89a80759f47c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.969500 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-etcd-service-ca\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.974011 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-etcd-ca\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.978880 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/658992cc-be88-49f4-b35c-bc445583fc39-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.991894 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/658992cc-be88-49f4-b35c-bc445583fc39-trusted-ca\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.992996 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/658992cc-be88-49f4-b35c-bc445583fc39-registry-certificates\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:10 crc kubenswrapper[4857]: I1201 21:37:10.996368 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-registry-tls\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.016395 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-serving-cert\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.016810 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/658992cc-be88-49f4-b35c-bc445583fc39-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122477 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc175021-8815-4fd7-b42c-af7e29b281fd-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mfv8h\" (UID: \"fc175021-8815-4fd7-b42c-af7e29b281fd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122526 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d668b26f-a9b6-4544-abf9-64e72728e342-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mxkzn\" (UID: \"d668b26f-a9b6-4544-abf9-64e72728e342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122552 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htc9q\" (UniqueName: \"kubernetes.io/projected/1cb66d76-6c0a-48c9-a0a8-54a60f7aae01-kube-api-access-htc9q\") pod \"control-plane-machine-set-operator-78cbb6b69f-8mmlp\" (UID: \"1cb66d76-6c0a-48c9-a0a8-54a60f7aae01\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122573 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/01903422-1e15-44cf-87f1-6037f6631592-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mbzs2\" (UID: \"01903422-1e15-44cf-87f1-6037f6631592\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122600 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2rr8\" (UniqueName: \"kubernetes.io/projected/1f215aad-96e0-4995-af49-99d7b947e674-kube-api-access-h2rr8\") pod \"ingress-canary-qd6km\" (UID: \"1f215aad-96e0-4995-af49-99d7b947e674\") " pod="openshift-ingress-canary/ingress-canary-qd6km" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122621 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq6hw\" (UniqueName: \"kubernetes.io/projected/6c8375eb-1803-475e-b897-9307d7bc29ee-kube-api-access-nq6hw\") pod \"collect-profiles-29410410-9gpfb\" (UID: \"6c8375eb-1803-475e-b897-9307d7bc29ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122641 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt57p\" (UniqueName: \"kubernetes.io/projected/6c4ff7e6-723d-4085-9bbc-bc865c8e6f61-kube-api-access-rt57p\") pod \"packageserver-d55dfcdfc-wg2k4\" (UID: \"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122661 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed50d226-e00b-447f-be13-e5847b1cf84b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6ls6r\" (UID: \"ed50d226-e00b-447f-be13-e5847b1cf84b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122685 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/555f0013-cb67-45f1-b96b-3f2b938fba11-service-ca-bundle\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122704 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-socket-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122731 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-csi-data-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122752 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc175021-8815-4fd7-b42c-af7e29b281fd-config\") pod \"kube-apiserver-operator-766d6c64bb-mfv8h\" (UID: \"fc175021-8815-4fd7-b42c-af7e29b281fd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122771 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f215aad-96e0-4995-af49-99d7b947e674-cert\") pod \"ingress-canary-qd6km\" (UID: \"1f215aad-96e0-4995-af49-99d7b947e674\") " pod="openshift-ingress-canary/ingress-canary-qd6km" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122790 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b1d82f9-f291-417b-bf3a-e8085c0a8f2b-serving-cert\") pod \"service-ca-operator-777779d784-jt8d4\" (UID: \"4b1d82f9-f291-417b-bf3a-e8085c0a8f2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122809 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhr5c\" (UniqueName: \"kubernetes.io/projected/e7f9d27b-4546-443e-aecc-d2478cb1a5ae-kube-api-access-vhr5c\") pod \"kube-storage-version-migrator-operator-b67b599dd-242k4\" (UID: \"e7f9d27b-4546-443e-aecc-d2478cb1a5ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122832 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6c4ff7e6-723d-4085-9bbc-bc865c8e6f61-tmpfs\") pod \"packageserver-d55dfcdfc-wg2k4\" (UID: \"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122855 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d668b26f-a9b6-4544-abf9-64e72728e342-proxy-tls\") pod \"machine-config-controller-84d6567774-mxkzn\" (UID: \"d668b26f-a9b6-4544-abf9-64e72728e342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122877 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7f9d27b-4546-443e-aecc-d2478cb1a5ae-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-242k4\" (UID: \"e7f9d27b-4546-443e-aecc-d2478cb1a5ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122898 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1d82f9-f291-417b-bf3a-e8085c0a8f2b-config\") pod \"service-ca-operator-777779d784-jt8d4\" (UID: \"4b1d82f9-f291-417b-bf3a-e8085c0a8f2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122930 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/1cb66d76-6c0a-48c9-a0a8-54a60f7aae01-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8mmlp\" (UID: \"1cb66d76-6c0a-48c9-a0a8-54a60f7aae01\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122952 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4da3afa4-fa1f-42e5-b321-d59419e58910-signing-key\") pod \"service-ca-9c57cc56f-nrp5f\" (UID: \"4da3afa4-fa1f-42e5-b321-d59419e58910\") " pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122975 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/25436d81-ab74-476f-ae9f-e74ca9452da0-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xrmc2\" (UID: \"25436d81-ab74-476f-ae9f-e74ca9452da0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.122994 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6c4ff7e6-723d-4085-9bbc-bc865c8e6f61-apiservice-cert\") pod \"packageserver-d55dfcdfc-wg2k4\" (UID: \"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.123017 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4lvj\" (UniqueName: \"kubernetes.io/projected/555f0013-cb67-45f1-b96b-3f2b938fba11-kube-api-access-d4lvj\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.123053 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nrtj9\" (UID: \"dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.123070 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0e090c12-a85e-4a29-9698-98e00f798808-images\") pod \"machine-config-operator-74547568cd-jp84x\" (UID: \"0e090c12-a85e-4a29-9698-98e00f798808\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.123090 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ng9h\" (UniqueName: \"kubernetes.io/projected/01903422-1e15-44cf-87f1-6037f6631592-kube-api-access-5ng9h\") pod \"marketplace-operator-79b997595-mbzs2\" (UID: \"01903422-1e15-44cf-87f1-6037f6631592\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.125988 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpjx6\" (UniqueName: \"kubernetes.io/projected/2e0f8071-e17c-4d6f-aefe-771448acbe1a-kube-api-access-mpjx6\") pod \"machine-config-server-p44n8\" (UID: \"2e0f8071-e17c-4d6f-aefe-771448acbe1a\") " pod="openshift-machine-config-operator/machine-config-server-p44n8" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126034 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/25436d81-ab74-476f-ae9f-e74ca9452da0-srv-cert\") pod \"olm-operator-6b444d44fb-xrmc2\" (UID: \"25436d81-ab74-476f-ae9f-e74ca9452da0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126079 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/555f0013-cb67-45f1-b96b-3f2b938fba11-stats-auth\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126097 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed50d226-e00b-447f-be13-e5847b1cf84b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6ls6r\" (UID: \"ed50d226-e00b-447f-be13-e5847b1cf84b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126120 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmdtc\" (UniqueName: \"kubernetes.io/projected/d668b26f-a9b6-4544-abf9-64e72728e342-kube-api-access-jmdtc\") pod \"machine-config-controller-84d6567774-mxkzn\" (UID: \"d668b26f-a9b6-4544-abf9-64e72728e342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126144 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clc47\" (UniqueName: \"kubernetes.io/projected/4da3afa4-fa1f-42e5-b321-d59419e58910-kube-api-access-clc47\") pod \"service-ca-9c57cc56f-nrp5f\" (UID: \"4da3afa4-fa1f-42e5-b321-d59419e58910\") " pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126167 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6590c60-f467-48a5-8551-0599bf6cb370-metrics-tls\") pod \"dns-default-tgwpg\" (UID: \"d6590c60-f467-48a5-8551-0599bf6cb370\") " pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126187 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/555f0013-cb67-45f1-b96b-3f2b938fba11-default-certificate\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126212 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nrtj9\" (UID: \"dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126238 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6c4ff7e6-723d-4085-9bbc-bc865c8e6f61-webhook-cert\") pod \"packageserver-d55dfcdfc-wg2k4\" (UID: \"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126271 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/511408a8-c761-46fe-bb8d-dd1147c0728e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-n5slz\" (UID: \"511408a8-c761-46fe-bb8d-dd1147c0728e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126298 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98czm\" (UniqueName: \"kubernetes.io/projected/25436d81-ab74-476f-ae9f-e74ca9452da0-kube-api-access-98czm\") pod \"olm-operator-6b444d44fb-xrmc2\" (UID: \"25436d81-ab74-476f-ae9f-e74ca9452da0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126321 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9zkf\" (UniqueName: \"kubernetes.io/projected/d6590c60-f467-48a5-8551-0599bf6cb370-kube-api-access-v9zkf\") pod \"dns-default-tgwpg\" (UID: \"d6590c60-f467-48a5-8551-0599bf6cb370\") " pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126346 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nrtj9\" (UID: \"dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126375 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-registration-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126399 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mk8s\" (UniqueName: \"kubernetes.io/projected/4b1d82f9-f291-417b-bf3a-e8085c0a8f2b-kube-api-access-9mk8s\") pod \"service-ca-operator-777779d784-jt8d4\" (UID: \"4b1d82f9-f291-417b-bf3a-e8085c0a8f2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126426 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/2e0f8071-e17c-4d6f-aefe-771448acbe1a-certs\") pod \"machine-config-server-p44n8\" (UID: \"2e0f8071-e17c-4d6f-aefe-771448acbe1a\") " pod="openshift-machine-config-operator/machine-config-server-p44n8" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126450 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/555f0013-cb67-45f1-b96b-3f2b938fba11-metrics-certs\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126468 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/2e0f8071-e17c-4d6f-aefe-771448acbe1a-node-bootstrap-token\") pod \"machine-config-server-p44n8\" (UID: \"2e0f8071-e17c-4d6f-aefe-771448acbe1a\") " pod="openshift-machine-config-operator/machine-config-server-p44n8" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126506 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed50d226-e00b-447f-be13-e5847b1cf84b-config\") pod \"kube-controller-manager-operator-78b949d7b-6ls6r\" (UID: \"ed50d226-e00b-447f-be13-e5847b1cf84b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126531 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xgds\" (UniqueName: \"kubernetes.io/projected/511408a8-c761-46fe-bb8d-dd1147c0728e-kube-api-access-2xgds\") pod \"package-server-manager-789f6589d5-n5slz\" (UID: \"511408a8-c761-46fe-bb8d-dd1147c0728e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126555 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4da3afa4-fa1f-42e5-b321-d59419e58910-signing-cabundle\") pod \"service-ca-9c57cc56f-nrp5f\" (UID: \"4da3afa4-fa1f-42e5-b321-d59419e58910\") " pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126573 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e780f3c5-76bd-47b9-97ff-89a80759f47c-profile-collector-cert\") pod \"catalog-operator-68c6474976-kkf64\" (UID: \"e780f3c5-76bd-47b9-97ff-89a80759f47c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126596 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0e090c12-a85e-4a29-9698-98e00f798808-proxy-tls\") pod \"machine-config-operator-74547568cd-jp84x\" (UID: \"0e090c12-a85e-4a29-9698-98e00f798808\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126617 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-plugins-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126638 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csj66\" (UniqueName: \"kubernetes.io/projected/e780f3c5-76bd-47b9-97ff-89a80759f47c-kube-api-access-csj66\") pod \"catalog-operator-68c6474976-kkf64\" (UID: \"e780f3c5-76bd-47b9-97ff-89a80759f47c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126666 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126688 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-mountpoint-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126705 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6590c60-f467-48a5-8551-0599bf6cb370-config-volume\") pod \"dns-default-tgwpg\" (UID: \"d6590c60-f467-48a5-8551-0599bf6cb370\") " pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126729 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01903422-1e15-44cf-87f1-6037f6631592-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mbzs2\" (UID: \"01903422-1e15-44cf-87f1-6037f6631592\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126754 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7f9d27b-4546-443e-aecc-d2478cb1a5ae-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-242k4\" (UID: \"e7f9d27b-4546-443e-aecc-d2478cb1a5ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126775 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e780f3c5-76bd-47b9-97ff-89a80759f47c-srv-cert\") pod \"catalog-operator-68c6474976-kkf64\" (UID: \"e780f3c5-76bd-47b9-97ff-89a80759f47c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126805 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc175021-8815-4fd7-b42c-af7e29b281fd-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mfv8h\" (UID: \"fc175021-8815-4fd7-b42c-af7e29b281fd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126826 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbpk9\" (UniqueName: \"kubernetes.io/projected/5361e52d-63b2-4f66-b786-9b1348a6b428-kube-api-access-kbpk9\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126841 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c8375eb-1803-475e-b897-9307d7bc29ee-secret-volume\") pod \"collect-profiles-29410410-9gpfb\" (UID: \"6c8375eb-1803-475e-b897-9307d7bc29ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126863 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98b9r\" (UniqueName: \"kubernetes.io/projected/0e090c12-a85e-4a29-9698-98e00f798808-kube-api-access-98b9r\") pod \"machine-config-operator-74547568cd-jp84x\" (UID: \"0e090c12-a85e-4a29-9698-98e00f798808\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126895 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0e090c12-a85e-4a29-9698-98e00f798808-auth-proxy-config\") pod \"machine-config-operator-74547568cd-jp84x\" (UID: \"0e090c12-a85e-4a29-9698-98e00f798808\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.126918 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c8375eb-1803-475e-b897-9307d7bc29ee-config-volume\") pod \"collect-profiles-29410410-9gpfb\" (UID: \"6c8375eb-1803-475e-b897-9307d7bc29ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.129028 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c8375eb-1803-475e-b897-9307d7bc29ee-config-volume\") pod \"collect-profiles-29410410-9gpfb\" (UID: \"6c8375eb-1803-475e-b897-9307d7bc29ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.134114 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7g4cw"] Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.145724 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7f9d27b-4546-443e-aecc-d2478cb1a5ae-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-242k4\" (UID: \"e7f9d27b-4546-443e-aecc-d2478cb1a5ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.146296 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0e090c12-a85e-4a29-9698-98e00f798808-images\") pod \"machine-config-operator-74547568cd-jp84x\" (UID: \"0e090c12-a85e-4a29-9698-98e00f798808\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.147818 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nrtj9\" (UID: \"dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.148304 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1d82f9-f291-417b-bf3a-e8085c0a8f2b-config\") pod \"service-ca-operator-777779d784-jt8d4\" (UID: \"4b1d82f9-f291-417b-bf3a-e8085c0a8f2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.149663 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d668b26f-a9b6-4544-abf9-64e72728e342-proxy-tls\") pod \"machine-config-controller-84d6567774-mxkzn\" (UID: \"d668b26f-a9b6-4544-abf9-64e72728e342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.150868 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/555f0013-cb67-45f1-b96b-3f2b938fba11-stats-auth\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:11 crc kubenswrapper[4857]: E1201 21:37:11.152155 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:11.652139114 +0000 UTC m=+150.142201431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.152305 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-mountpoint-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.152461 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/25436d81-ab74-476f-ae9f-e74ca9452da0-srv-cert\") pod \"olm-operator-6b444d44fb-xrmc2\" (UID: \"25436d81-ab74-476f-ae9f-e74ca9452da0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.156452 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6590c60-f467-48a5-8551-0599bf6cb370-config-volume\") pod \"dns-default-tgwpg\" (UID: \"d6590c60-f467-48a5-8551-0599bf6cb370\") " pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.156576 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d668b26f-a9b6-4544-abf9-64e72728e342-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mxkzn\" (UID: \"d668b26f-a9b6-4544-abf9-64e72728e342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.161759 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clc47\" (UniqueName: \"kubernetes.io/projected/4da3afa4-fa1f-42e5-b321-d59419e58910-kube-api-access-clc47\") pod \"service-ca-9c57cc56f-nrp5f\" (UID: \"4da3afa4-fa1f-42e5-b321-d59419e58910\") " pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.162294 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6c4ff7e6-723d-4085-9bbc-bc865c8e6f61-apiservice-cert\") pod \"packageserver-d55dfcdfc-wg2k4\" (UID: \"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.162470 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6c4ff7e6-723d-4085-9bbc-bc865c8e6f61-tmpfs\") pod \"packageserver-d55dfcdfc-wg2k4\" (UID: \"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.163968 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4da3afa4-fa1f-42e5-b321-d59419e58910-signing-key\") pod \"service-ca-9c57cc56f-nrp5f\" (UID: \"4da3afa4-fa1f-42e5-b321-d59419e58910\") " pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.164913 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7f9d27b-4546-443e-aecc-d2478cb1a5ae-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-242k4\" (UID: \"e7f9d27b-4546-443e-aecc-d2478cb1a5ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.168444 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-csi-data-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.168631 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-socket-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.168662 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/555f0013-cb67-45f1-b96b-3f2b938fba11-service-ca-bundle\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.169163 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed50d226-e00b-447f-be13-e5847b1cf84b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6ls6r\" (UID: \"ed50d226-e00b-447f-be13-e5847b1cf84b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.169186 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc175021-8815-4fd7-b42c-af7e29b281fd-config\") pod \"kube-apiserver-operator-766d6c64bb-mfv8h\" (UID: \"fc175021-8815-4fd7-b42c-af7e29b281fd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.171032 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nrtj9\" (UID: \"dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.171221 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-registration-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.172386 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/1cb66d76-6c0a-48c9-a0a8-54a60f7aae01-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8mmlp\" (UID: \"1cb66d76-6c0a-48c9-a0a8-54a60f7aae01\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.176931 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01903422-1e15-44cf-87f1-6037f6631592-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mbzs2\" (UID: \"01903422-1e15-44cf-87f1-6037f6631592\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.178809 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/25436d81-ab74-476f-ae9f-e74ca9452da0-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xrmc2\" (UID: \"25436d81-ab74-476f-ae9f-e74ca9452da0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.178911 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0e090c12-a85e-4a29-9698-98e00f798808-auth-proxy-config\") pod \"machine-config-operator-74547568cd-jp84x\" (UID: \"0e090c12-a85e-4a29-9698-98e00f798808\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.179484 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/01903422-1e15-44cf-87f1-6037f6631592-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mbzs2\" (UID: \"01903422-1e15-44cf-87f1-6037f6631592\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.179623 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed50d226-e00b-447f-be13-e5847b1cf84b-config\") pod \"kube-controller-manager-operator-78b949d7b-6ls6r\" (UID: \"ed50d226-e00b-447f-be13-e5847b1cf84b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.180586 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5361e52d-63b2-4f66-b786-9b1348a6b428-plugins-dir\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.180859 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b1d82f9-f291-417b-bf3a-e8085c0a8f2b-serving-cert\") pod \"service-ca-operator-777779d784-jt8d4\" (UID: \"4b1d82f9-f291-417b-bf3a-e8085c0a8f2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.181403 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4da3afa4-fa1f-42e5-b321-d59419e58910-signing-cabundle\") pod \"service-ca-9c57cc56f-nrp5f\" (UID: \"4da3afa4-fa1f-42e5-b321-d59419e58910\") " pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.186871 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xghgp\" (UniqueName: \"kubernetes.io/projected/3301b37b-d60f-42d2-91dd-c0c898019831-kube-api-access-xghgp\") pod \"multus-admission-controller-857f4d67dd-rg5g2\" (UID: \"3301b37b-d60f-42d2-91dd-c0c898019831\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rg5g2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.186897 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qnkc\" (UniqueName: \"kubernetes.io/projected/e3c0a9de-bb5f-44d5-b99f-f7acd1d0d308-kube-api-access-9qnkc\") pod \"migrator-59844c95c7-6875l\" (UID: \"e3c0a9de-bb5f-44d5-b99f-f7acd1d0d308\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6875l" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.187546 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-bound-sa-token\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.187556 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h69dr\" (UniqueName: \"kubernetes.io/projected/5f801a62-b8a7-4fd8-af84-5b3ca694af6d-kube-api-access-h69dr\") pod \"etcd-operator-b45778765-4mvhq\" (UID: \"5f801a62-b8a7-4fd8-af84-5b3ca694af6d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.187948 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ng9h\" (UniqueName: \"kubernetes.io/projected/01903422-1e15-44cf-87f1-6037f6631592-kube-api-access-5ng9h\") pod \"marketplace-operator-79b997595-mbzs2\" (UID: \"01903422-1e15-44cf-87f1-6037f6631592\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.188712 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e780f3c5-76bd-47b9-97ff-89a80759f47c-srv-cert\") pod \"catalog-operator-68c6474976-kkf64\" (UID: \"e780f3c5-76bd-47b9-97ff-89a80759f47c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.189343 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4lvj\" (UniqueName: \"kubernetes.io/projected/555f0013-cb67-45f1-b96b-3f2b938fba11-kube-api-access-d4lvj\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.190540 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/511408a8-c761-46fe-bb8d-dd1147c0728e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-n5slz\" (UID: \"511408a8-c761-46fe-bb8d-dd1147c0728e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.203020 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp9pd\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-kube-api-access-zp9pd\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.213104 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c8375eb-1803-475e-b897-9307d7bc29ee-secret-volume\") pod \"collect-profiles-29410410-9gpfb\" (UID: \"6c8375eb-1803-475e-b897-9307d7bc29ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.215226 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2rr8\" (UniqueName: \"kubernetes.io/projected/1f215aad-96e0-4995-af49-99d7b947e674-kube-api-access-h2rr8\") pod \"ingress-canary-qd6km\" (UID: \"1f215aad-96e0-4995-af49-99d7b947e674\") " pod="openshift-ingress-canary/ingress-canary-qd6km" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.215241 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc175021-8815-4fd7-b42c-af7e29b281fd-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mfv8h\" (UID: \"fc175021-8815-4fd7-b42c-af7e29b281fd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.215808 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6c4ff7e6-723d-4085-9bbc-bc865c8e6f61-webhook-cert\") pod \"packageserver-d55dfcdfc-wg2k4\" (UID: \"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.215830 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq6hw\" (UniqueName: \"kubernetes.io/projected/6c8375eb-1803-475e-b897-9307d7bc29ee-kube-api-access-nq6hw\") pod \"collect-profiles-29410410-9gpfb\" (UID: \"6c8375eb-1803-475e-b897-9307d7bc29ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.216605 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f215aad-96e0-4995-af49-99d7b947e674-cert\") pod \"ingress-canary-qd6km\" (UID: \"1f215aad-96e0-4995-af49-99d7b947e674\") " pod="openshift-ingress-canary/ingress-canary-qd6km" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.216834 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/2e0f8071-e17c-4d6f-aefe-771448acbe1a-node-bootstrap-token\") pod \"machine-config-server-p44n8\" (UID: \"2e0f8071-e17c-4d6f-aefe-771448acbe1a\") " pod="openshift-machine-config-operator/machine-config-server-p44n8" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.217349 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/2e0f8071-e17c-4d6f-aefe-771448acbe1a-certs\") pod \"machine-config-server-p44n8\" (UID: \"2e0f8071-e17c-4d6f-aefe-771448acbe1a\") " pod="openshift-machine-config-operator/machine-config-server-p44n8" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.223824 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt57p\" (UniqueName: \"kubernetes.io/projected/6c4ff7e6-723d-4085-9bbc-bc865c8e6f61-kube-api-access-rt57p\") pod \"packageserver-d55dfcdfc-wg2k4\" (UID: \"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.223932 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpjx6\" (UniqueName: \"kubernetes.io/projected/2e0f8071-e17c-4d6f-aefe-771448acbe1a-kube-api-access-mpjx6\") pod \"machine-config-server-p44n8\" (UID: \"2e0f8071-e17c-4d6f-aefe-771448acbe1a\") " pod="openshift-machine-config-operator/machine-config-server-p44n8" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.224128 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/555f0013-cb67-45f1-b96b-3f2b938fba11-metrics-certs\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.224316 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/555f0013-cb67-45f1-b96b-3f2b938fba11-default-certificate\") pod \"router-default-5444994796-z26dc\" (UID: \"555f0013-cb67-45f1-b96b-3f2b938fba11\") " pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.224338 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6590c60-f467-48a5-8551-0599bf6cb370-metrics-tls\") pod \"dns-default-tgwpg\" (UID: \"d6590c60-f467-48a5-8551-0599bf6cb370\") " pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.224506 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.224530 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbpk9\" (UniqueName: \"kubernetes.io/projected/5361e52d-63b2-4f66-b786-9b1348a6b428-kube-api-access-kbpk9\") pod \"csi-hostpathplugin-6mll7\" (UID: \"5361e52d-63b2-4f66-b786-9b1348a6b428\") " pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.227629 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:11 crc kubenswrapper[4857]: E1201 21:37:11.228235 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:11.728217396 +0000 UTC m=+150.218279713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.228908 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e780f3c5-76bd-47b9-97ff-89a80759f47c-profile-collector-cert\") pod \"catalog-operator-68c6474976-kkf64\" (UID: \"e780f3c5-76bd-47b9-97ff-89a80759f47c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.229387 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0e090c12-a85e-4a29-9698-98e00f798808-proxy-tls\") pod \"machine-config-operator-74547568cd-jp84x\" (UID: \"0e090c12-a85e-4a29-9698-98e00f798808\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.235255 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.235523 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc175021-8815-4fd7-b42c-af7e29b281fd-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mfv8h\" (UID: \"fc175021-8815-4fd7-b42c-af7e29b281fd\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" Dec 01 21:37:11 crc kubenswrapper[4857]: E1201 21:37:11.235709 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:11.735691831 +0000 UTC m=+150.225754148 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.256193 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.317733 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed50d226-e00b-447f-be13-e5847b1cf84b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6ls6r\" (UID: \"ed50d226-e00b-447f-be13-e5847b1cf84b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.320802 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.322144 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xgds\" (UniqueName: \"kubernetes.io/projected/511408a8-c761-46fe-bb8d-dd1147c0728e-kube-api-access-2xgds\") pod \"package-server-manager-789f6589d5-n5slz\" (UID: \"511408a8-c761-46fe-bb8d-dd1147c0728e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.341908 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.348387 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-p44n8" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.434592 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.435152 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9zkf\" (UniqueName: \"kubernetes.io/projected/d6590c60-f467-48a5-8551-0599bf6cb370-kube-api-access-v9zkf\") pod \"dns-default-tgwpg\" (UID: \"d6590c60-f467-48a5-8551-0599bf6cb370\") " pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.435916 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qd6km" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.440366 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98czm\" (UniqueName: \"kubernetes.io/projected/25436d81-ab74-476f-ae9f-e74ca9452da0-kube-api-access-98czm\") pod \"olm-operator-6b444d44fb-xrmc2\" (UID: \"25436d81-ab74-476f-ae9f-e74ca9452da0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.444560 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mk8s\" (UniqueName: \"kubernetes.io/projected/4b1d82f9-f291-417b-bf3a-e8085c0a8f2b-kube-api-access-9mk8s\") pod \"service-ca-operator-777779d784-jt8d4\" (UID: \"4b1d82f9-f291-417b-bf3a-e8085c0a8f2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.447173 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhr5c\" (UniqueName: \"kubernetes.io/projected/e7f9d27b-4546-443e-aecc-d2478cb1a5ae-kube-api-access-vhr5c\") pod \"kube-storage-version-migrator-operator-b67b599dd-242k4\" (UID: \"e7f9d27b-4546-443e-aecc-d2478cb1a5ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.450151 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nrtj9\" (UID: \"dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.451175 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.451678 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:11 crc kubenswrapper[4857]: E1201 21:37:11.452228 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:11.952210929 +0000 UTC m=+150.442273246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.452308 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.452621 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6mll7" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.452889 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6875l" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.453117 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rg5g2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.453351 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.453664 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htc9q\" (UniqueName: \"kubernetes.io/projected/1cb66d76-6c0a-48c9-a0a8-54a60f7aae01-kube-api-access-htc9q\") pod \"control-plane-machine-set-operator-78cbb6b69f-8mmlp\" (UID: \"1cb66d76-6c0a-48c9-a0a8-54a60f7aae01\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.453777 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.454570 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.454612 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.455590 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmdtc\" (UniqueName: \"kubernetes.io/projected/d668b26f-a9b6-4544-abf9-64e72728e342-kube-api-access-jmdtc\") pod \"machine-config-controller-84d6567774-mxkzn\" (UID: \"d668b26f-a9b6-4544-abf9-64e72728e342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.461134 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.479874 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.486947 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98b9r\" (UniqueName: \"kubernetes.io/projected/0e090c12-a85e-4a29-9698-98e00f798808-kube-api-access-98b9r\") pod \"machine-config-operator-74547568cd-jp84x\" (UID: \"0e090c12-a85e-4a29-9698-98e00f798808\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.489128 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csj66\" (UniqueName: \"kubernetes.io/projected/e780f3c5-76bd-47b9-97ff-89a80759f47c-kube-api-access-csj66\") pod \"catalog-operator-68c6474976-kkf64\" (UID: \"e780f3c5-76bd-47b9-97ff-89a80759f47c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.504214 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.586919 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:11 crc kubenswrapper[4857]: E1201 21:37:11.587219 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:12.087207644 +0000 UTC m=+150.577269961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.587536 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.629271 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.687545 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:11 crc kubenswrapper[4857]: E1201 21:37:11.688495 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:12.188472326 +0000 UTC m=+150.678534643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.717312 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.813629 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:11 crc kubenswrapper[4857]: E1201 21:37:11.813911 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:12.31389785 +0000 UTC m=+150.803960167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.814416 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.878217 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" event={"ID":"5952edee-93a8-4b8c-9fa0-e15699b86f5d","Type":"ContainerStarted","Data":"f8a1bb27954517af34823be668cfb02f4e605f18dd9f32dd41b3cc8f801491d5"} Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.878573 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" event={"ID":"369ee7eb-c09c-4406-90de-5c3a4442d0c5","Type":"ContainerStarted","Data":"53b99fde9c16a2dbf7ee63d423d730664f93b78ce30a77fa70e6340bd06e0126"} Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.891736 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-p44n8" event={"ID":"2e0f8071-e17c-4d6f-aefe-771448acbe1a","Type":"ContainerStarted","Data":"0bbbf8824991f3c27c2ddeac70b410b56747cba418d64dca2360be2b3682a483"} Dec 01 21:37:11 crc kubenswrapper[4857]: I1201 21:37:11.914611 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:11 crc kubenswrapper[4857]: E1201 21:37:11.915179 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:12.415157542 +0000 UTC m=+150.905219859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:11.932123 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7g4cw" event={"ID":"17e47df6-fb87-44b9-a1a8-143d629f900a","Type":"ContainerStarted","Data":"7d150d86f6d9027eeaa7d97b15f6eb205665abdbdd34d46ed8551afb29cc8674"} Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.019002 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.019756 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:12.519738965 +0000 UTC m=+151.009801272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.123809 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.124054 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:12.62402673 +0000 UTC m=+151.114089037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.124156 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.124964 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:12.624956066 +0000 UTC m=+151.115018373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.181669 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.192085 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p"] Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.287967 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.288185 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:12.788152073 +0000 UTC m=+151.278214390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.288555 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.288903 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:12.788896394 +0000 UTC m=+151.278958711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.313153 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mmb9s"] Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.377344 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2mhm7"] Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.389881 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.390488 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:12.890440323 +0000 UTC m=+151.380502650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.407493 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s"] Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.411995 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mmv46" podStartSLOduration=132.411978653 podStartE2EDuration="2m12.411978653s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:12.410878003 +0000 UTC m=+150.900940320" watchObservedRunningTime="2025-12-01 21:37:12.411978653 +0000 UTC m=+150.902040970" Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.491120 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.491477 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:12.991464949 +0000 UTC m=+151.481527266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.576201 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" podStartSLOduration=132.576184698 podStartE2EDuration="2m12.576184698s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:12.572868337 +0000 UTC m=+151.062930654" watchObservedRunningTime="2025-12-01 21:37:12.576184698 +0000 UTC m=+151.066247015" Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.592703 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.592854 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.092837064 +0000 UTC m=+151.582899381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.592936 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.593356 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.093342748 +0000 UTC m=+151.583405075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: W1201 21:37:12.630366 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod334d54e0_4413_43e2_875d_bbca874149be.slice/crio-eb101ca81e1b4e5a862843407353135d5d130a01400ed8609c8f651857dfad8a WatchSource:0}: Error finding container eb101ca81e1b4e5a862843407353135d5d130a01400ed8609c8f651857dfad8a: Status 404 returned error can't find the container with id eb101ca81e1b4e5a862843407353135d5d130a01400ed8609c8f651857dfad8a Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.656161 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-r5kwb" podStartSLOduration=132.656138707 podStartE2EDuration="2m12.656138707s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:12.650446501 +0000 UTC m=+151.140508838" watchObservedRunningTime="2025-12-01 21:37:12.656138707 +0000 UTC m=+151.146201014" Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.676404 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8"] Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.691108 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-w54ht"] Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.694421 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.694583 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.194554398 +0000 UTC m=+151.684616715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.694768 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.695162 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.195141265 +0000 UTC m=+151.685203582 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.796290 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.796421 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.296395097 +0000 UTC m=+151.786457414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.796659 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.797146 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.297130367 +0000 UTC m=+151.787192674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.897689 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:12 crc kubenswrapper[4857]: E1201 21:37:12.898476 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.398450371 +0000 UTC m=+151.888512678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.940635 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" event={"ID":"7ff6afde-262c-4ebf-be69-2e043814271c","Type":"ContainerStarted","Data":"3e2a1dc94974b96496e9e77a178d31736c515dc08b50c015a61647b4267cda17"} Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.942050 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7g4cw" event={"ID":"17e47df6-fb87-44b9-a1a8-143d629f900a","Type":"ContainerStarted","Data":"d175d1be6d88b4b70df12d82257df9135bdd9e6cf9408b4d1cd75a3921c2f4fa"} Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.947179 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" event={"ID":"334d54e0-4413-43e2-875d-bbca874149be","Type":"ContainerStarted","Data":"eb101ca81e1b4e5a862843407353135d5d130a01400ed8609c8f651857dfad8a"} Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.955772 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2mhm7" event={"ID":"d26df1fc-905c-4a0e-8933-fabccc185cbc","Type":"ContainerStarted","Data":"08ec10807138ffe4ad6df2adedb47870bac3174191bba33ecfbc1a54dc178d44"} Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.963176 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" event={"ID":"1927b12d-f48b-4ac4-b61a-ca2cc610312b","Type":"ContainerStarted","Data":"63d529c504becfdca9b8a459d8c3fcf14c1a13e9b973fb62eb6d4e6a55172631"} Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.968394 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-pcsb9" podStartSLOduration=132.968358174 podStartE2EDuration="2m12.968358174s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:12.967958973 +0000 UTC m=+151.458021310" watchObservedRunningTime="2025-12-01 21:37:12.968358174 +0000 UTC m=+151.458420691" Dec 01 21:37:12 crc kubenswrapper[4857]: I1201 21:37:12.969228 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-z26dc" event={"ID":"555f0013-cb67-45f1-b96b-3f2b938fba11","Type":"ContainerStarted","Data":"702515a9af15f39b2de754431111861a7e8e72ec42d580a7f9c859351d7dc4ca"} Dec 01 21:37:13 crc kubenswrapper[4857]: I1201 21:37:13.000388 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:13 crc kubenswrapper[4857]: E1201 21:37:13.000909 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.500888255 +0000 UTC m=+151.990950582 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:13 crc kubenswrapper[4857]: I1201 21:37:13.101535 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:13 crc kubenswrapper[4857]: E1201 21:37:13.101948 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.601929681 +0000 UTC m=+152.091991998 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:13 crc kubenswrapper[4857]: I1201 21:37:13.106341 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rxh8n" podStartSLOduration=133.106322291 podStartE2EDuration="2m13.106322291s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:13.089586703 +0000 UTC m=+151.579649020" watchObservedRunningTime="2025-12-01 21:37:13.106322291 +0000 UTC m=+151.596384608" Dec 01 21:37:13 crc kubenswrapper[4857]: I1201 21:37:13.202563 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:13 crc kubenswrapper[4857]: E1201 21:37:13.203197 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.703165042 +0000 UTC m=+152.193227559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:13 crc kubenswrapper[4857]: W1201 21:37:13.210385 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f264c9f_c454_4521_a501_6e230287a919.slice/crio-10177b09e6a1402dceaf781a088ecbbc510e1854c43dd67b02f66b59bd59e99c WatchSource:0}: Error finding container 10177b09e6a1402dceaf781a088ecbbc510e1854c43dd67b02f66b59bd59e99c: Status 404 returned error can't find the container with id 10177b09e6a1402dceaf781a088ecbbc510e1854c43dd67b02f66b59bd59e99c Dec 01 21:37:13 crc kubenswrapper[4857]: I1201 21:37:13.303833 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:13 crc kubenswrapper[4857]: E1201 21:37:13.304023 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.803989712 +0000 UTC m=+152.294052039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:13 crc kubenswrapper[4857]: I1201 21:37:13.304504 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:13 crc kubenswrapper[4857]: E1201 21:37:13.304882 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.804868416 +0000 UTC m=+152.294930733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:13 crc kubenswrapper[4857]: I1201 21:37:13.405200 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:13 crc kubenswrapper[4857]: E1201 21:37:13.405481 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:13.90546677 +0000 UTC m=+152.395529087 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:13 crc kubenswrapper[4857]: I1201 21:37:13.506126 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:13 crc kubenswrapper[4857]: E1201 21:37:13.506545 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:14.006526097 +0000 UTC m=+152.496588414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:13 crc kubenswrapper[4857]: I1201 21:37:13.635632 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:13 crc kubenswrapper[4857]: E1201 21:37:13.636024 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:14.136008982 +0000 UTC m=+152.626071299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:13 crc kubenswrapper[4857]: I1201 21:37:13.736983 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:13 crc kubenswrapper[4857]: E1201 21:37:13.737532 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:14.237519891 +0000 UTC m=+152.727582198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:13 crc kubenswrapper[4857]: I1201 21:37:13.838339 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:13 crc kubenswrapper[4857]: E1201 21:37:13.838570 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:14.338555426 +0000 UTC m=+152.828617743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:13 crc kubenswrapper[4857]: I1201 21:37:13.939281 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:13 crc kubenswrapper[4857]: E1201 21:37:13.939545 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:14.43953198 +0000 UTC m=+152.929594287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.020984 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" event={"ID":"1927b12d-f48b-4ac4-b61a-ca2cc610312b","Type":"ContainerStarted","Data":"74f01568da1db67f45db94ea649f5a3767510771186c8cba39efad40821eee19"} Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.021822 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-g6p59"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.024431 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-p44n8" event={"ID":"2e0f8071-e17c-4d6f-aefe-771448acbe1a","Type":"ContainerStarted","Data":"4cfc00c8933a1108a8d1fd36069a8383c6bc848318e4798abc4813131bde86c6"} Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.041062 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" event={"ID":"6f264c9f-c454-4521-a501-6e230287a919","Type":"ContainerStarted","Data":"10177b09e6a1402dceaf781a088ecbbc510e1854c43dd67b02f66b59bd59e99c"} Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.042761 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:14 crc kubenswrapper[4857]: E1201 21:37:14.044125 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:14.544095312 +0000 UTC m=+153.034157629 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.044953 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-z26dc" event={"ID":"555f0013-cb67-45f1-b96b-3f2b938fba11","Type":"ContainerStarted","Data":"b241fd628cb65bbd8e06215525248140e51317dd7dd65ae4f0df3de8e02ad00a"} Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.047225 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:14 crc kubenswrapper[4857]: E1201 21:37:14.048467 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:14.548447921 +0000 UTC m=+153.038510238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.048852 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" event={"ID":"7ff6afde-262c-4ebf-be69-2e043814271c","Type":"ContainerStarted","Data":"0cdc776a88afcca486adf56ecc6ec9b4d2b0cd60649a605fdf9766f0524322cd"} Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.049078 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.050657 4857 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mmb9s container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.050928 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" podUID="7ff6afde-262c-4ebf-be69-2e043814271c" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.052732 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7g4cw" event={"ID":"17e47df6-fb87-44b9-a1a8-143d629f900a","Type":"ContainerStarted","Data":"ba3fd60d5c8a81935d4c1f2ddd0acbd5e5dda68f9923c5e7d859c55d0cf8d8cb"} Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.056092 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" event={"ID":"334d54e0-4413-43e2-875d-bbca874149be","Type":"ContainerStarted","Data":"de6cd4690b9a52e095ca4801898bf3079e5dc787dc15f1e0d212449fba95e667"} Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.060840 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2mhm7" event={"ID":"d26df1fc-905c-4a0e-8933-fabccc185cbc","Type":"ContainerStarted","Data":"59162994bd880e8434f45d324131c27d3413c16b287a58f8f1922d2a3cf1b27b"} Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.061406 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.063424 4857 patch_prober.go:28] interesting pod/console-operator-58897d9998-2mhm7 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.063457 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-2mhm7" podUID="d26df1fc-905c-4a0e-8933-fabccc185cbc" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.066524 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-w54ht" event={"ID":"0fd59b67-77d3-4e4f-b0b4-4bc714e29e73","Type":"ContainerStarted","Data":"d1560349e655fd8c9a7d9e06611584afe5ad0dac79a502ca1f365d5649b93add"} Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.066553 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-w54ht" event={"ID":"0fd59b67-77d3-4e4f-b0b4-4bc714e29e73","Type":"ContainerStarted","Data":"cd2af97cc5ad472358c38dd70c5dc0fa1aa8525572a666438f05938b7c27e05f"} Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.066571 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-w54ht" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.073563 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-w54ht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.073620 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w54ht" podUID="0fd59b67-77d3-4e4f-b0b4-4bc714e29e73" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.073759 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wgv2p" podStartSLOduration=134.073749034 podStartE2EDuration="2m14.073749034s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:14.067551214 +0000 UTC m=+152.557613531" watchObservedRunningTime="2025-12-01 21:37:14.073749034 +0000 UTC m=+152.563811351" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.079195 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7x54g" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.109319 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-2mhm7" podStartSLOduration=134.109285557 podStartE2EDuration="2m14.109285557s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:14.087609643 +0000 UTC m=+152.577671960" watchObservedRunningTime="2025-12-01 21:37:14.109285557 +0000 UTC m=+152.599347874" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.126617 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-7g4cw" podStartSLOduration=134.126599841 podStartE2EDuration="2m14.126599841s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:14.120631327 +0000 UTC m=+152.610693654" watchObservedRunningTime="2025-12-01 21:37:14.126599841 +0000 UTC m=+152.616662158" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.148415 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:14 crc kubenswrapper[4857]: E1201 21:37:14.149695 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:14.649678122 +0000 UTC m=+153.139740439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.155407 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-w54ht" podStartSLOduration=134.155386229 podStartE2EDuration="2m14.155386229s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:14.15434778 +0000 UTC m=+152.644410097" watchObservedRunningTime="2025-12-01 21:37:14.155386229 +0000 UTC m=+152.645448546" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.221034 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-z26dc" podStartSLOduration=134.221013555 podStartE2EDuration="2m14.221013555s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:14.210675372 +0000 UTC m=+152.700737699" watchObservedRunningTime="2025-12-01 21:37:14.221013555 +0000 UTC m=+152.711075872" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.252600 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:14 crc kubenswrapper[4857]: E1201 21:37:14.253020 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:14.753002871 +0000 UTC m=+153.243065188 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.276262 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" podStartSLOduration=134.276245537 podStartE2EDuration="2m14.276245537s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:14.250435331 +0000 UTC m=+152.740497648" watchObservedRunningTime="2025-12-01 21:37:14.276245537 +0000 UTC m=+152.766307854" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.279646 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-p44n8" podStartSLOduration=6.27963623 podStartE2EDuration="6.27963623s" podCreationTimestamp="2025-12-01 21:37:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:14.275345402 +0000 UTC m=+152.765407719" watchObservedRunningTime="2025-12-01 21:37:14.27963623 +0000 UTC m=+152.769698547" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.292716 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6875l"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.353979 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:14 crc kubenswrapper[4857]: E1201 21:37:14.354469 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:14.854452028 +0000 UTC m=+153.344514345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.375222 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qd6km"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.406230 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nrp5f"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.408866 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.424141 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.428404 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.431357 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.454185 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.456962 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:14 crc kubenswrapper[4857]: E1201 21:37:14.457302 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:14.957288983 +0000 UTC m=+153.447351300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.473885 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.479585 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-2bwxl"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.480059 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:14 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:14 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:14 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.480096 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.482392 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5wggv"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.491113 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4mvhq"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.494419 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mbzs2"] Dec 01 21:37:14 crc kubenswrapper[4857]: W1201 21:37:14.500447 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c2b2eb0_e49a_4ce5_ab44_d1abe7f45842.slice/crio-eeb448ec0d042245947919bf11e16e36b6d57548ba1d0ec7f313c0be30ef7a8a WatchSource:0}: Error finding container eeb448ec0d042245947919bf11e16e36b6d57548ba1d0ec7f313c0be30ef7a8a: Status 404 returned error can't find the container with id eeb448ec0d042245947919bf11e16e36b6d57548ba1d0ec7f313c0be30ef7a8a Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.502799 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.509839 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.522951 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.535955 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6mll7"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.557840 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:14 crc kubenswrapper[4857]: E1201 21:37:14.557975 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:15.057940499 +0000 UTC m=+153.548002816 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.558030 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:14 crc kubenswrapper[4857]: E1201 21:37:14.558499 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:15.058479363 +0000 UTC m=+153.548541680 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.559118 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.563723 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rg5g2"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.565492 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.573662 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.626728 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.659901 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:14 crc kubenswrapper[4857]: E1201 21:37:14.661491 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:15.161472553 +0000 UTC m=+153.651534870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.662530 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb"] Dec 01 21:37:14 crc kubenswrapper[4857]: W1201 21:37:14.681146 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cb66d76_6c0a_48c9_a0a8_54a60f7aae01.slice/crio-45078251fef2b729493374769cd4b60deee151fdabcb88b9287b3362685488d7 WatchSource:0}: Error finding container 45078251fef2b729493374769cd4b60deee151fdabcb88b9287b3362685488d7: Status 404 returned error can't find the container with id 45078251fef2b729493374769cd4b60deee151fdabcb88b9287b3362685488d7 Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.713058 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.762266 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-tgwpg"] Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.762680 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:14 crc kubenswrapper[4857]: E1201 21:37:14.762970 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:15.262957351 +0000 UTC m=+153.753019668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.783367 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4"] Dec 01 21:37:14 crc kubenswrapper[4857]: W1201 21:37:14.838522 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded50d226_e00b_447f_be13_e5847b1cf84b.slice/crio-665a4376bb67c61f98269182e2737b621c31ae7bed31b4e89406aba8b48f451d WatchSource:0}: Error finding container 665a4376bb67c61f98269182e2737b621c31ae7bed31b4e89406aba8b48f451d: Status 404 returned error can't find the container with id 665a4376bb67c61f98269182e2737b621c31ae7bed31b4e89406aba8b48f451d Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.863374 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:14 crc kubenswrapper[4857]: E1201 21:37:14.864066 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:15.364017188 +0000 UTC m=+153.854079505 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:14 crc kubenswrapper[4857]: I1201 21:37:14.965110 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:14 crc kubenswrapper[4857]: E1201 21:37:14.965464 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:15.465452535 +0000 UTC m=+153.955514842 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.100744 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:15 crc kubenswrapper[4857]: E1201 21:37:15.101487 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:15.601461318 +0000 UTC m=+154.091523635 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.101972 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" event={"ID":"369ee7eb-c09c-4406-90de-5c3a4442d0c5","Type":"ContainerStarted","Data":"601029ff39d11be6b0733391f8aa66a9e00e11fe0821735f2cb8cb737634f8dd"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.138311 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fs9lv" podStartSLOduration=135.138297776 podStartE2EDuration="2m15.138297776s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:15.137836084 +0000 UTC m=+153.627898411" watchObservedRunningTime="2025-12-01 21:37:15.138297776 +0000 UTC m=+153.628360093" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.143132 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" event={"ID":"4da3afa4-fa1f-42e5-b321-d59419e58910","Type":"ContainerStarted","Data":"acbe9efb9e8a399519b2adf1ef7431df8f44d646ed69294863cb23d2906474e5"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.175836 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" event={"ID":"dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1","Type":"ContainerStarted","Data":"ad22d71f1785e4a37229b59b9ee4b9a843caf08f2a722076e907aea6da9946e8"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.184986 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" event={"ID":"5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842","Type":"ContainerStarted","Data":"eeb448ec0d042245947919bf11e16e36b6d57548ba1d0ec7f313c0be30ef7a8a"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.199834 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" event={"ID":"71476481-b2d8-4a26-b8a8-66fa5ec7135e","Type":"ContainerStarted","Data":"7b8c6257dd9710830ff496d98c9652e5d63a25ba837c0fed7aacc6d9fbf5be7f"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.199887 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" event={"ID":"71476481-b2d8-4a26-b8a8-66fa5ec7135e","Type":"ContainerStarted","Data":"c14c303356ee459aa88762763e0742a5185b1d073917334d9529fb24976065cb"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.200518 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.203643 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:15 crc kubenswrapper[4857]: E1201 21:37:15.205395 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:15.705380833 +0000 UTC m=+154.195443150 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.205836 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" event={"ID":"fc175021-8815-4fd7-b42c-af7e29b281fd","Type":"ContainerStarted","Data":"5ab0322f4be4342d6749dc75ebd8c833c66af01e5d0150487fedc091dc66bac5"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.212149 4857 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-5wggv container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.212191 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" podUID="71476481-b2d8-4a26-b8a8-66fa5ec7135e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.228202 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" event={"ID":"d668b26f-a9b6-4544-abf9-64e72728e342","Type":"ContainerStarted","Data":"1042314a460133ee301dbc7512bd4c418d30218ce91dc91676ce5fa0062d4bbd"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.245732 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" event={"ID":"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61","Type":"ContainerStarted","Data":"927cc31b506ef8d4b174de85766240628b386b57a19504e1c52699b470a96d44"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.245773 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" event={"ID":"6c4ff7e6-723d-4085-9bbc-bc865c8e6f61","Type":"ContainerStarted","Data":"dea8538b85d747a82f01c2cab7b2ee9d327a780d9797c2fa8b9b7ae11788bbe3"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.248195 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.249257 4857 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wg2k4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.249326 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" podUID="6c4ff7e6-723d-4085-9bbc-bc865c8e6f61" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.250656 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp" event={"ID":"1cb66d76-6c0a-48c9-a0a8-54a60f7aae01","Type":"ContainerStarted","Data":"45078251fef2b729493374769cd4b60deee151fdabcb88b9287b3362685488d7"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.285911 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" podStartSLOduration=135.285896647 podStartE2EDuration="2m15.285896647s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:15.232656579 +0000 UTC m=+153.722718896" watchObservedRunningTime="2025-12-01 21:37:15.285896647 +0000 UTC m=+153.775958964" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.305658 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:15 crc kubenswrapper[4857]: E1201 21:37:15.305966 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:15.805945956 +0000 UTC m=+154.296008273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.306370 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:15 crc kubenswrapper[4857]: E1201 21:37:15.307544 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:15.807530859 +0000 UTC m=+154.297593176 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.314370 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" event={"ID":"ed50d226-e00b-447f-be13-e5847b1cf84b","Type":"ContainerStarted","Data":"665a4376bb67c61f98269182e2737b621c31ae7bed31b4e89406aba8b48f451d"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.328071 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" event={"ID":"5f801a62-b8a7-4fd8-af84-5b3ca694af6d","Type":"ContainerStarted","Data":"b05925c777032b50e737449a86bbf5bf48f6aeef5eb52add2e28c5e434abd485"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.365164 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" event={"ID":"e780f3c5-76bd-47b9-97ff-89a80759f47c","Type":"ContainerStarted","Data":"35e085d4a9cc481bcaad5bf4a88f50e07ad3cb4fd1e127faaec7d6530eab7b63"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.407597 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:15 crc kubenswrapper[4857]: E1201 21:37:15.408054 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:15.90802671 +0000 UTC m=+154.398089027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.413076 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rg5g2" event={"ID":"3301b37b-d60f-42d2-91dd-c0c898019831","Type":"ContainerStarted","Data":"d01c5f3804f641f412d005efdbfdb561ccd62de33903038534ae900ac4a44888"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.432169 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" event={"ID":"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4","Type":"ContainerStarted","Data":"1abadb1083c5d04c0f67cd729e853b51bc799f90e7111b1ec7928d9ce0d59265"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.475473 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" event={"ID":"4b1d82f9-f291-417b-bf3a-e8085c0a8f2b","Type":"ContainerStarted","Data":"b74733b89aa512bcef4f77709ab34524931de9a81d404a0e01410ec8598c483a"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.475641 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:15 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:15 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:15 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.475670 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.549714 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:15 crc kubenswrapper[4857]: E1201 21:37:15.550023 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:16.050009967 +0000 UTC m=+154.540072284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.558765 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" event={"ID":"0e090c12-a85e-4a29-9698-98e00f798808","Type":"ContainerStarted","Data":"c509341f617a2c43ee1ed7fcf1c1e24f931b05accc6b0a46233534fa7486584c"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.560267 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" event={"ID":"01903422-1e15-44cf-87f1-6037f6631592","Type":"ContainerStarted","Data":"e73b46dff6eae12c378780cdbf27e9ebb112abe51ac617a3b1d30b5516d06391"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.561225 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6mll7" event={"ID":"5361e52d-63b2-4f66-b786-9b1348a6b428","Type":"ContainerStarted","Data":"53accf062eb8a42dff45de695be21f4a4eab9d0529c32c7f7c9c643dc5dceaf7"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.562079 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" event={"ID":"25436d81-ab74-476f-ae9f-e74ca9452da0","Type":"ContainerStarted","Data":"dac934875452070991e5336ba9bcdec9708244092c0dbffbe72850b8da47c1e7"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.563009 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6875l" event={"ID":"e3c0a9de-bb5f-44d5-b99f-f7acd1d0d308","Type":"ContainerStarted","Data":"163173f9d65860975138cb9f9ef23b150b888227d5fc825d7ae8f387db2c613d"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.563055 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6875l" event={"ID":"e3c0a9de-bb5f-44d5-b99f-f7acd1d0d308","Type":"ContainerStarted","Data":"f38882cc711417064ade5e32acd78b988de879a0b11a18cd5d4a5e8830977d68"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.563676 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" event={"ID":"511408a8-c761-46fe-bb8d-dd1147c0728e","Type":"ContainerStarted","Data":"4b86e7bc1badf7cc11f73fd56f76540a8a44d2b1f9bbd08e71bc13d0f42251f7"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.565199 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" event={"ID":"6f264c9f-c454-4521-a501-6e230287a919","Type":"ContainerStarted","Data":"ea609cedcf687f86981971fdfcf2abf17eca7943c7219039df9d4cdc862b4315"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.565222 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" event={"ID":"6f264c9f-c454-4521-a501-6e230287a919","Type":"ContainerStarted","Data":"961c57dea35f493875555602f7dddca1dff3a62f7cd4ebc4b3c58ee09e16fbbd"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.567399 4857 generic.go:334] "Generic (PLEG): container finished" podID="b6911089-a8c9-4e07-8f85-5d5d2ff9f736" containerID="b53dd1670c314a4e112e9a66d95f619068951c82cb7b6f0b59d810c7e7917cf6" exitCode=0 Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.567456 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-g6p59" event={"ID":"b6911089-a8c9-4e07-8f85-5d5d2ff9f736","Type":"ContainerDied","Data":"b53dd1670c314a4e112e9a66d95f619068951c82cb7b6f0b59d810c7e7917cf6"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.567473 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-g6p59" event={"ID":"b6911089-a8c9-4e07-8f85-5d5d2ff9f736","Type":"ContainerStarted","Data":"8d9c6227ccf51872a501a820451efc391af10c33e55b5cb88f942a2080389b66"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.568822 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" event={"ID":"6c8375eb-1803-475e-b897-9307d7bc29ee","Type":"ContainerStarted","Data":"8cab48f9ff506e8fff1cab5058b3294002b336b6eea39817dae02ac5a5cb37c7"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.570170 4857 generic.go:334] "Generic (PLEG): container finished" podID="334d54e0-4413-43e2-875d-bbca874149be" containerID="de6cd4690b9a52e095ca4801898bf3079e5dc787dc15f1e0d212449fba95e667" exitCode=0 Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.570450 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" event={"ID":"334d54e0-4413-43e2-875d-bbca874149be","Type":"ContainerDied","Data":"de6cd4690b9a52e095ca4801898bf3079e5dc787dc15f1e0d212449fba95e667"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.570604 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" event={"ID":"334d54e0-4413-43e2-875d-bbca874149be","Type":"ContainerStarted","Data":"59257f89fa66d01cd2316e32856502cf83dc2956852122354e7bd9d5430baacf"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.577021 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" event={"ID":"e7f9d27b-4546-443e-aecc-d2478cb1a5ae","Type":"ContainerStarted","Data":"e34c485eaf599aa251cdf306d77c2427473970376594c23ed1efbad916159428"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.584847 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qd6km" event={"ID":"1f215aad-96e0-4995-af49-99d7b947e674","Type":"ContainerStarted","Data":"8dfac17753cf6cf1a560d9c6be2d78d9d4a7b22c22411f171aa5c06ef8cf05c2"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.584889 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qd6km" event={"ID":"1f215aad-96e0-4995-af49-99d7b947e674","Type":"ContainerStarted","Data":"59e4bbb69576f6d3f8b0f474709fa0163ca30ea3bd3810253c4de27b53b62904"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.588939 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" podStartSLOduration=134.588926823 podStartE2EDuration="2m14.588926823s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:15.328638747 +0000 UTC m=+153.818701064" watchObservedRunningTime="2025-12-01 21:37:15.588926823 +0000 UTC m=+154.078989130" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.591358 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" event={"ID":"5c55767b-b626-4e74-a0fe-23d71cf0e347","Type":"ContainerStarted","Data":"f62e1960bd921e45a702c7c7ff0fb33a3ebcd178f24e33c43a196123c36f0814"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.591446 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" event={"ID":"5c55767b-b626-4e74-a0fe-23d71cf0e347","Type":"ContainerStarted","Data":"e3efe52763acd8d35ae2e0299a2af5806dfebd75e6b4797580ee885b8f138b44"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.592168 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.611869 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tgwpg" event={"ID":"d6590c60-f467-48a5-8551-0599bf6cb370","Type":"ContainerStarted","Data":"397aa80ec37b7108f4c547867fdb35f8f2af6176235a07f033ff8ed7fc23a086"} Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.612216 4857 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-8rjjr container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.614504 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" podUID="5c55767b-b626-4e74-a0fe-23d71cf0e347" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.613921 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-w54ht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.614722 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w54ht" podUID="0fd59b67-77d3-4e4f-b0b4-4bc714e29e73" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.622767 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqzb8" podStartSLOduration=135.622757109 podStartE2EDuration="2m15.622757109s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:15.591094462 +0000 UTC m=+154.081156769" watchObservedRunningTime="2025-12-01 21:37:15.622757109 +0000 UTC m=+154.112819426" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.636378 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.639448 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-2mhm7" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.651617 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:15 crc kubenswrapper[4857]: E1201 21:37:15.651801 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:16.151775003 +0000 UTC m=+154.641837320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.651883 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.652133 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.652252 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:15 crc kubenswrapper[4857]: E1201 21:37:15.667506 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:16.167487103 +0000 UTC m=+154.657549420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.705669 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-qd6km" podStartSLOduration=7.705656048 podStartE2EDuration="7.705656048s" podCreationTimestamp="2025-12-01 21:37:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:15.662687842 +0000 UTC m=+154.152750159" watchObservedRunningTime="2025-12-01 21:37:15.705656048 +0000 UTC m=+154.195718365" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.733478 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" podStartSLOduration=134.733462749 podStartE2EDuration="2m14.733462749s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:15.713240766 +0000 UTC m=+154.203303083" watchObservedRunningTime="2025-12-01 21:37:15.733462749 +0000 UTC m=+154.223525056" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.780396 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" podStartSLOduration=134.780375034 podStartE2EDuration="2m14.780375034s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:15.770517254 +0000 UTC m=+154.260579571" watchObservedRunningTime="2025-12-01 21:37:15.780375034 +0000 UTC m=+154.270437351" Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.788259 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:15 crc kubenswrapper[4857]: E1201 21:37:15.788862 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:16.288847586 +0000 UTC m=+154.778909903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.897594 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:15 crc kubenswrapper[4857]: E1201 21:37:15.898022 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:16.398006444 +0000 UTC m=+154.888068751 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:15 crc kubenswrapper[4857]: I1201 21:37:15.999574 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:15 crc kubenswrapper[4857]: E1201 21:37:15.999924 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:16.499907044 +0000 UTC m=+154.989969361 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.016357 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.104268 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:16 crc kubenswrapper[4857]: E1201 21:37:16.104698 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:16.604683202 +0000 UTC m=+155.094745519 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.239716 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:16 crc kubenswrapper[4857]: E1201 21:37:16.240013 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:16.739998356 +0000 UTC m=+155.230060673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.381647 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:16 crc kubenswrapper[4857]: E1201 21:37:16.382147 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:16.882130917 +0000 UTC m=+155.372193234 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.477593 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:16 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:16 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:16 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.477955 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.491179 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:16 crc kubenswrapper[4857]: E1201 21:37:16.491909 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:16.991882552 +0000 UTC m=+155.481944869 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.592547 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:16 crc kubenswrapper[4857]: E1201 21:37:16.592986 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:17.092973129 +0000 UTC m=+155.583035446 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.694083 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:16 crc kubenswrapper[4857]: E1201 21:37:16.694595 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:17.19455852 +0000 UTC m=+155.684620837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.712849 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" event={"ID":"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4","Type":"ContainerStarted","Data":"d47e7b39cdb10550392cd0d08ed33b7e7e0e1e3f2937c4f2a5df7213e14cca30"} Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.728769 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" event={"ID":"4b1d82f9-f291-417b-bf3a-e8085c0a8f2b","Type":"ContainerStarted","Data":"fe3391af7d0c8f61f4ed490140dacadd434cc8b4e2d4deececb1ae80dd1761ad"} Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.764598 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jt8d4" podStartSLOduration=135.764581167 podStartE2EDuration="2m15.764581167s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:16.763858648 +0000 UTC m=+155.253920965" watchObservedRunningTime="2025-12-01 21:37:16.764581167 +0000 UTC m=+155.254643474" Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.765336 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6875l" event={"ID":"e3c0a9de-bb5f-44d5-b99f-f7acd1d0d308","Type":"ContainerStarted","Data":"eec22044f85950319cd18eacf3d98cd2543f4931c12eca1ee759ab9c32453e9b"} Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.800707 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:16 crc kubenswrapper[4857]: E1201 21:37:16.801032 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:17.301020315 +0000 UTC m=+155.791082632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.805956 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" event={"ID":"01903422-1e15-44cf-87f1-6037f6631592","Type":"ContainerStarted","Data":"9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a"} Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.807235 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.821488 4857 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mbzs2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.821535 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" podUID="01903422-1e15-44cf-87f1-6037f6631592" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.822416 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6875l" podStartSLOduration=135.82240127 podStartE2EDuration="2m15.82240127s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:16.821206948 +0000 UTC m=+155.311269265" watchObservedRunningTime="2025-12-01 21:37:16.82240127 +0000 UTC m=+155.312463587" Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.849088 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" event={"ID":"e780f3c5-76bd-47b9-97ff-89a80759f47c","Type":"ContainerStarted","Data":"11a290bd477a895838fca2a9e36e32a6aa135cf126ea3f4ef54287fdf6781e7d"} Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.850328 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.852359 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" podStartSLOduration=135.85233031 podStartE2EDuration="2m15.85233031s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:16.850815998 +0000 UTC m=+155.340878315" watchObservedRunningTime="2025-12-01 21:37:16.85233031 +0000 UTC m=+155.342392627" Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.866189 4857 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-kkf64 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.866252 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" podUID="e780f3c5-76bd-47b9-97ff-89a80759f47c" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.871672 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rg5g2" event={"ID":"3301b37b-d60f-42d2-91dd-c0c898019831","Type":"ContainerStarted","Data":"7ca5216669b2e17262b27b93e4ff9348080826058745f78e5d2c7b104939f88b"} Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.873910 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" event={"ID":"dd329494-d7ca-4bf2-bcc2-6d6301e3a5d1","Type":"ContainerStarted","Data":"d08edb3b072ed4b7b3f2044f27a99c20054a7f6ef123833b1e9839dc88705e08"} Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.894299 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" podStartSLOduration=135.894282098 podStartE2EDuration="2m15.894282098s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:16.893916418 +0000 UTC m=+155.383978735" watchObservedRunningTime="2025-12-01 21:37:16.894282098 +0000 UTC m=+155.384344415" Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.900671 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" event={"ID":"fc175021-8815-4fd7-b42c-af7e29b281fd","Type":"ContainerStarted","Data":"d7181bc39efcf7af273a87e42fe7a67f2f63f68df327772a0efa2ffc5c168cd9"} Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.901276 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:16 crc kubenswrapper[4857]: E1201 21:37:16.902829 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:17.402804831 +0000 UTC m=+155.892867148 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.913953 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tgwpg" event={"ID":"d6590c60-f467-48a5-8551-0599bf6cb370","Type":"ContainerStarted","Data":"35359ef4b5161afd05d0ed504c52a6b0f8f2b072de7bcdc4f55aee9b1d006fda"} Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.916448 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nrtj9" podStartSLOduration=135.916433564 podStartE2EDuration="2m15.916433564s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:16.914610094 +0000 UTC m=+155.404672411" watchObservedRunningTime="2025-12-01 21:37:16.916433564 +0000 UTC m=+155.406495881" Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.978858 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mfv8h" podStartSLOduration=135.978840713 podStartE2EDuration="2m15.978840713s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:16.976605842 +0000 UTC m=+155.466668159" watchObservedRunningTime="2025-12-01 21:37:16.978840713 +0000 UTC m=+155.468903030" Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.988297 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" event={"ID":"0e090c12-a85e-4a29-9698-98e00f798808","Type":"ContainerStarted","Data":"33dcd09af2664d5afb2cd82355c4d7ab4e804ef11c38f3580ba5f1e7819a3ac1"} Dec 01 21:37:16 crc kubenswrapper[4857]: I1201 21:37:16.988403 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" event={"ID":"0e090c12-a85e-4a29-9698-98e00f798808","Type":"ContainerStarted","Data":"77d7328bf606d7b50f32e982be96ddf0501189c7fd02568bbcd34ce7129f49e0"} Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.004586 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.005838 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:17.505827922 +0000 UTC m=+155.995890239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.007336 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" event={"ID":"d668b26f-a9b6-4544-abf9-64e72728e342","Type":"ContainerStarted","Data":"86133537159db42989a2c84921f895f8c97584e4ab3dff3e9c37a96b191c5209"} Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.062587 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" event={"ID":"4da3afa4-fa1f-42e5-b321-d59419e58910","Type":"ContainerStarted","Data":"fa79e3205cd9d08fccd1672eae81fd3396a01b4b624fd79a539095b1147a56f7"} Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.065358 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jp84x" podStartSLOduration=136.065333311 podStartE2EDuration="2m16.065333311s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:17.061825175 +0000 UTC m=+155.551887492" watchObservedRunningTime="2025-12-01 21:37:17.065333311 +0000 UTC m=+155.555395628" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.075425 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp" event={"ID":"1cb66d76-6c0a-48c9-a0a8-54a60f7aae01","Type":"ContainerStarted","Data":"47f2c3d43432c5046543f4b57b14437523dd2dedc007fcd21b870601689b43f2"} Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.092933 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" event={"ID":"ed50d226-e00b-447f-be13-e5847b1cf84b","Type":"ContainerStarted","Data":"41ef147e1ff31ce09a06a8714adefb3f346b3844d0b7c259f7d2d8e0959efa5b"} Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.107688 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.108342 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:17.608316807 +0000 UTC m=+156.098379124 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.127459 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ls6r" podStartSLOduration=136.127441571 podStartE2EDuration="2m16.127441571s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:17.126276979 +0000 UTC m=+155.616339316" watchObservedRunningTime="2025-12-01 21:37:17.127441571 +0000 UTC m=+155.617503888" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.129459 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" podStartSLOduration=136.129453306 podStartE2EDuration="2m16.129453306s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:17.096896645 +0000 UTC m=+155.586958962" watchObservedRunningTime="2025-12-01 21:37:17.129453306 +0000 UTC m=+155.619515623" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.170010 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8mmlp" podStartSLOduration=136.169991456 podStartE2EDuration="2m16.169991456s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:17.169849242 +0000 UTC m=+155.659911559" watchObservedRunningTime="2025-12-01 21:37:17.169991456 +0000 UTC m=+155.660053773" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.207211 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" event={"ID":"25436d81-ab74-476f-ae9f-e74ca9452da0","Type":"ContainerStarted","Data":"8a670c4c1fa9da5ca1104bee7d092f1093598d3ddbf4e88c0f6337377c300b0e"} Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.207414 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.208918 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.211321 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:17.711309876 +0000 UTC m=+156.201372193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.212696 4857 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-xrmc2 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.212756 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" podUID="25436d81-ab74-476f-ae9f-e74ca9452da0" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.215415 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-nrp5f" podStartSLOduration=136.215401639 podStartE2EDuration="2m16.215401639s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:17.213050694 +0000 UTC m=+155.703113011" watchObservedRunningTime="2025-12-01 21:37:17.215401639 +0000 UTC m=+155.705463956" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.249778 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" event={"ID":"5f801a62-b8a7-4fd8-af84-5b3ca694af6d","Type":"ContainerStarted","Data":"268ec3e7824372efad278db0dfae0d4d71a2e10f592b8e7a4c25633c3e8a1504"} Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.265218 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" podStartSLOduration=136.265198272 podStartE2EDuration="2m16.265198272s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:17.235370965 +0000 UTC m=+155.725433282" watchObservedRunningTime="2025-12-01 21:37:17.265198272 +0000 UTC m=+155.755260589" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.266679 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" podStartSLOduration=136.266674822 podStartE2EDuration="2m16.266674822s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:17.2651379 +0000 UTC m=+155.755200217" watchObservedRunningTime="2025-12-01 21:37:17.266674822 +0000 UTC m=+155.756737139" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.276742 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" event={"ID":"5c2b2eb0-e49a-4ce5-ab44-d1abe7f45842","Type":"ContainerStarted","Data":"bb1a4ed1a47adb778994859f8bb67378fe6f3345313dde85239617267b44c2b3"} Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.314914 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.316297 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.317895 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:17.817855253 +0000 UTC m=+156.307917570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.319271 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.320091 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:17.820063814 +0000 UTC m=+156.310126131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.326613 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wdc4s" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.382833 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-4mvhq" podStartSLOduration=137.382781161 podStartE2EDuration="2m17.382781161s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:17.295718337 +0000 UTC m=+155.785780654" watchObservedRunningTime="2025-12-01 21:37:17.382781161 +0000 UTC m=+155.872843478" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.388191 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pwq6n" podStartSLOduration=137.388182249 podStartE2EDuration="2m17.388182249s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:17.314592474 +0000 UTC m=+155.804654801" watchObservedRunningTime="2025-12-01 21:37:17.388182249 +0000 UTC m=+155.878244566" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.486988 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.488536 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:17.988520036 +0000 UTC m=+156.478582353 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.511642 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.563660 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:17 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:17 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:17 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.563726 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.589570 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.590070 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:18.090052574 +0000 UTC m=+156.580114891 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.690503 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.690993 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:18.190971567 +0000 UTC m=+156.681033884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.792550 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.792967 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:18.292948848 +0000 UTC m=+156.783011165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.828868 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wg2k4" Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.894203 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.894408 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:18.394374015 +0000 UTC m=+156.884436322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.894456 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.894839 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:18.394826207 +0000 UTC m=+156.884888524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.996000 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.996403 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:18.496376617 +0000 UTC m=+156.986438934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:17 crc kubenswrapper[4857]: I1201 21:37:17.996454 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:17 crc kubenswrapper[4857]: E1201 21:37:17.996746 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:18.496733997 +0000 UTC m=+156.986796314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.097619 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:18 crc kubenswrapper[4857]: E1201 21:37:18.097969 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:18.597955228 +0000 UTC m=+157.088017545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.199484 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:18 crc kubenswrapper[4857]: E1201 21:37:18.200023 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:18.700001012 +0000 UTC m=+157.190063329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.300355 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:18 crc kubenswrapper[4857]: E1201 21:37:18.301051 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:18.801024147 +0000 UTC m=+157.291086464 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.335762 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rg5g2" event={"ID":"3301b37b-d60f-42d2-91dd-c0c898019831","Type":"ContainerStarted","Data":"cc1dc3b7d53e68b0ea11ae6585af5c50562558275dad06958224e98ef05f2523"} Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.337838 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6mll7" event={"ID":"5361e52d-63b2-4f66-b786-9b1348a6b428","Type":"ContainerStarted","Data":"ee03ddde70640ddfdd3014034c95a9e8068bc7804e06560145b683e396649bfd"} Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.339712 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-242k4" event={"ID":"e7f9d27b-4546-443e-aecc-d2478cb1a5ae","Type":"ContainerStarted","Data":"fdbdabd93e03aa065cb5f4c1fd4518b0b6e680411fa10ec60dc4e49605ef0fc5"} Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.345825 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mxkzn" event={"ID":"d668b26f-a9b6-4544-abf9-64e72728e342","Type":"ContainerStarted","Data":"13ae1655fd29642c75239885db93befa5acc727831eaca1a0b3694a4d48cab32"} Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.348106 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" event={"ID":"511408a8-c761-46fe-bb8d-dd1147c0728e","Type":"ContainerStarted","Data":"ef2179e75be00d756de1defd41e9e0fd9cf64f574b7d07a0c96501b55482a449"} Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.348142 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" event={"ID":"511408a8-c761-46fe-bb8d-dd1147c0728e","Type":"ContainerStarted","Data":"2b8bc817e4c447ff40491e60d8f66b005606179f7d461e7218fd26ef99639144"} Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.348217 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.350727 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tgwpg" event={"ID":"d6590c60-f467-48a5-8551-0599bf6cb370","Type":"ContainerStarted","Data":"46542e78298269f56bcbf1114fc7775b5bc039898151d83a791bec7665455f4e"} Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.350837 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.352423 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" event={"ID":"6c8375eb-1803-475e-b897-9307d7bc29ee","Type":"ContainerStarted","Data":"fdc6a1bc8c1714b9add47cbd440bbc801690cba6bde6e6b375970539e22fdff7"} Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.356696 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-g6p59" event={"ID":"b6911089-a8c9-4e07-8f85-5d5d2ff9f736","Type":"ContainerStarted","Data":"2d207f541c589c6838327bfcbfdc814887aca745777b9e4dc9b374b0279949ae"} Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.356803 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-g6p59" event={"ID":"b6911089-a8c9-4e07-8f85-5d5d2ff9f736","Type":"ContainerStarted","Data":"4ad61838be49f7e57606b1ef3e8c7046ec2a9f6b784d753d653eaf97a9db90b7"} Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.359381 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" event={"ID":"00d4b52b-5372-48a1-a1b4-6c17e18cc7a4","Type":"ContainerStarted","Data":"77d9a78c0822b3742392492880192f5cba770d85487a0fe3c0a8003e0b516358"} Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.360548 4857 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mbzs2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.360587 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" podUID="01903422-1e15-44cf-87f1-6037f6631592" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.361175 4857 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-xrmc2 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.361226 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" podUID="25436d81-ab74-476f-ae9f-e74ca9452da0" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.373308 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kkf64" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.402683 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:18 crc kubenswrapper[4857]: E1201 21:37:18.411704 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:18.911686767 +0000 UTC m=+157.401749084 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.515946 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:18 crc kubenswrapper[4857]: E1201 21:37:18.516475 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:19.016453055 +0000 UTC m=+157.506515372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.542381 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:18 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:18 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:18 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.542458 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.589945 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-rg5g2" podStartSLOduration=137.589918106 podStartE2EDuration="2m17.589918106s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:18.441823162 +0000 UTC m=+156.931885479" watchObservedRunningTime="2025-12-01 21:37:18.589918106 +0000 UTC m=+157.079980423" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.640593 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:18 crc kubenswrapper[4857]: E1201 21:37:18.640983 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:19.140970024 +0000 UTC m=+157.631032341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.672338 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" podStartSLOduration=137.672321142 podStartE2EDuration="2m17.672321142s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:18.671328465 +0000 UTC m=+157.161390772" watchObservedRunningTime="2025-12-01 21:37:18.672321142 +0000 UTC m=+157.162383459" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.683855 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" podStartSLOduration=138.681180204 podStartE2EDuration="2m18.681180204s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:18.59154385 +0000 UTC m=+157.081606167" watchObservedRunningTime="2025-12-01 21:37:18.681180204 +0000 UTC m=+157.171242521" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.743180 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:18 crc kubenswrapper[4857]: E1201 21:37:18.743780 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:19.243755017 +0000 UTC m=+157.733817344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.776592 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-g6p59" podStartSLOduration=138.776571766 podStartE2EDuration="2m18.776571766s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:18.773574334 +0000 UTC m=+157.263636661" watchObservedRunningTime="2025-12-01 21:37:18.776571766 +0000 UTC m=+157.266634083" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.874173 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:18 crc kubenswrapper[4857]: E1201 21:37:18.874643 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:19.37462614 +0000 UTC m=+157.864688457 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.968399 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-2bwxl" podStartSLOduration=137.968368706 podStartE2EDuration="2m17.968368706s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:18.808429808 +0000 UTC m=+157.298492115" watchObservedRunningTime="2025-12-01 21:37:18.968368706 +0000 UTC m=+157.458431023" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.976732 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:18 crc kubenswrapper[4857]: E1201 21:37:18.977146 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:19.477127486 +0000 UTC m=+157.967189803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.990846 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.991530 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-tgwpg" podStartSLOduration=10.99150649 podStartE2EDuration="10.99150649s" podCreationTimestamp="2025-12-01 21:37:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:18.970075363 +0000 UTC m=+157.460137670" watchObservedRunningTime="2025-12-01 21:37:18.99150649 +0000 UTC m=+157.481568807" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.992248 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.993244 4857 patch_prober.go:28] interesting pod/console-f9d7485db-r5kwb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.993296 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-r5kwb" podUID="b8e5475f-0369-448a-895e-c02b78884677" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.995629 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4s5mw"] Dec 01 21:37:18 crc kubenswrapper[4857]: I1201 21:37:18.997236 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.003604 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.078927 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szw7m\" (UniqueName: \"kubernetes.io/projected/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-kube-api-access-szw7m\") pod \"certified-operators-4s5mw\" (UID: \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\") " pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.079060 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-utilities\") pod \"certified-operators-4s5mw\" (UID: \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\") " pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.079728 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-catalog-content\") pod \"certified-operators-4s5mw\" (UID: \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\") " pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.079979 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.081169 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:19.581143644 +0000 UTC m=+158.071205961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.096338 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4s5mw"] Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.146907 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k9szd"] Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.148158 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:37:19 crc kubenswrapper[4857]: W1201 21:37:19.152289 4857 reflector.go:561] object-"openshift-marketplace"/"community-operators-dockercfg-dmngl": failed to list *v1.Secret: secrets "community-operators-dockercfg-dmngl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.152345 4857 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"community-operators-dockercfg-dmngl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"community-operators-dockercfg-dmngl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.181276 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.181516 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:19.68147743 +0000 UTC m=+158.171539747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.181641 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8049f37f-5448-4796-b384-691649bd559c-catalog-content\") pod \"community-operators-k9szd\" (UID: \"8049f37f-5448-4796-b384-691649bd559c\") " pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.181802 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szw7m\" (UniqueName: \"kubernetes.io/projected/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-kube-api-access-szw7m\") pod \"certified-operators-4s5mw\" (UID: \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\") " pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.181899 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-utilities\") pod \"certified-operators-4s5mw\" (UID: \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\") " pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.181940 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-catalog-content\") pod \"certified-operators-4s5mw\" (UID: \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\") " pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.181962 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phc6k\" (UniqueName: \"kubernetes.io/projected/8049f37f-5448-4796-b384-691649bd559c-kube-api-access-phc6k\") pod \"community-operators-k9szd\" (UID: \"8049f37f-5448-4796-b384-691649bd559c\") " pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.182194 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8049f37f-5448-4796-b384-691649bd559c-utilities\") pod \"community-operators-k9szd\" (UID: \"8049f37f-5448-4796-b384-691649bd559c\") " pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.182368 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-utilities\") pod \"certified-operators-4s5mw\" (UID: \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\") " pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.182709 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-catalog-content\") pod \"certified-operators-4s5mw\" (UID: \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\") " pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.234896 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k9szd"] Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.243145 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szw7m\" (UniqueName: \"kubernetes.io/projected/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-kube-api-access-szw7m\") pod \"certified-operators-4s5mw\" (UID: \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\") " pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.284254 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phc6k\" (UniqueName: \"kubernetes.io/projected/8049f37f-5448-4796-b384-691649bd559c-kube-api-access-phc6k\") pod \"community-operators-k9szd\" (UID: \"8049f37f-5448-4796-b384-691649bd559c\") " pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.284302 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8049f37f-5448-4796-b384-691649bd559c-utilities\") pod \"community-operators-k9szd\" (UID: \"8049f37f-5448-4796-b384-691649bd559c\") " pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.284336 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.284403 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8049f37f-5448-4796-b384-691649bd559c-catalog-content\") pod \"community-operators-k9szd\" (UID: \"8049f37f-5448-4796-b384-691649bd559c\") " pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.284895 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8049f37f-5448-4796-b384-691649bd559c-catalog-content\") pod \"community-operators-k9szd\" (UID: \"8049f37f-5448-4796-b384-691649bd559c\") " pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.285483 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8049f37f-5448-4796-b384-691649bd559c-utilities\") pod \"community-operators-k9szd\" (UID: \"8049f37f-5448-4796-b384-691649bd559c\") " pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.285831 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:19.785816697 +0000 UTC m=+158.275879004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.343645 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phc6k\" (UniqueName: \"kubernetes.io/projected/8049f37f-5448-4796-b384-691649bd559c-kube-api-access-phc6k\") pod \"community-operators-k9szd\" (UID: \"8049f37f-5448-4796-b384-691649bd559c\") " pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.370600 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6mll7" event={"ID":"5361e52d-63b2-4f66-b786-9b1348a6b428","Type":"ContainerStarted","Data":"ca893219ec67bdca9b413d257ea9f1d07a6df137e11d62288840117b61633c34"} Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.371835 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r2rmn"] Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.373379 4857 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mbzs2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.373440 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" podUID="01903422-1e15-44cf-87f1-6037f6631592" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.373884 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.385139 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.385306 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:19.88528105 +0000 UTC m=+158.375343367 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.385459 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.385794 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:19.885780843 +0000 UTC m=+158.375843150 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.391923 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r2rmn"] Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.403856 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.468523 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:19 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:19 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:19 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.468586 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.487263 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.487384 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:19.987357404 +0000 UTC m=+158.477419721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.487877 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.488118 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41882b8a-22d5-4283-a5e1-8a400181d9de-catalog-content\") pod \"certified-operators-r2rmn\" (UID: \"41882b8a-22d5-4283-a5e1-8a400181d9de\") " pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.488470 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2htwv\" (UniqueName: \"kubernetes.io/projected/41882b8a-22d5-4283-a5e1-8a400181d9de-kube-api-access-2htwv\") pod \"certified-operators-r2rmn\" (UID: \"41882b8a-22d5-4283-a5e1-8a400181d9de\") " pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.488557 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41882b8a-22d5-4283-a5e1-8a400181d9de-utilities\") pod \"certified-operators-r2rmn\" (UID: \"41882b8a-22d5-4283-a5e1-8a400181d9de\") " pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.489811 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:19.989790501 +0000 UTC m=+158.479852818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.538999 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7x8v8"] Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.540674 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.559571 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7x8v8"] Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.616382 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.616613 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0390146f-bbac-4efb-aaef-c7e9c41f8d51-catalog-content\") pod \"community-operators-7x8v8\" (UID: \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\") " pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.616685 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0390146f-bbac-4efb-aaef-c7e9c41f8d51-utilities\") pod \"community-operators-7x8v8\" (UID: \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\") " pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.616723 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2htwv\" (UniqueName: \"kubernetes.io/projected/41882b8a-22d5-4283-a5e1-8a400181d9de-kube-api-access-2htwv\") pod \"certified-operators-r2rmn\" (UID: \"41882b8a-22d5-4283-a5e1-8a400181d9de\") " pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.616751 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41882b8a-22d5-4283-a5e1-8a400181d9de-utilities\") pod \"certified-operators-r2rmn\" (UID: \"41882b8a-22d5-4283-a5e1-8a400181d9de\") " pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.616854 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41882b8a-22d5-4283-a5e1-8a400181d9de-catalog-content\") pod \"certified-operators-r2rmn\" (UID: \"41882b8a-22d5-4283-a5e1-8a400181d9de\") " pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.616887 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6zrm\" (UniqueName: \"kubernetes.io/projected/0390146f-bbac-4efb-aaef-c7e9c41f8d51-kube-api-access-q6zrm\") pod \"community-operators-7x8v8\" (UID: \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\") " pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.618332 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41882b8a-22d5-4283-a5e1-8a400181d9de-utilities\") pod \"certified-operators-r2rmn\" (UID: \"41882b8a-22d5-4283-a5e1-8a400181d9de\") " pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.618565 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41882b8a-22d5-4283-a5e1-8a400181d9de-catalog-content\") pod \"certified-operators-r2rmn\" (UID: \"41882b8a-22d5-4283-a5e1-8a400181d9de\") " pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.618630 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:20.118608827 +0000 UTC m=+158.608671224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.668320 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2htwv\" (UniqueName: \"kubernetes.io/projected/41882b8a-22d5-4283-a5e1-8a400181d9de-kube-api-access-2htwv\") pod \"certified-operators-r2rmn\" (UID: \"41882b8a-22d5-4283-a5e1-8a400181d9de\") " pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.691422 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.717815 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6zrm\" (UniqueName: \"kubernetes.io/projected/0390146f-bbac-4efb-aaef-c7e9c41f8d51-kube-api-access-q6zrm\") pod \"community-operators-7x8v8\" (UID: \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\") " pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.717891 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0390146f-bbac-4efb-aaef-c7e9c41f8d51-catalog-content\") pod \"community-operators-7x8v8\" (UID: \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\") " pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.717921 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0390146f-bbac-4efb-aaef-c7e9c41f8d51-utilities\") pod \"community-operators-7x8v8\" (UID: \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\") " pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.717962 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.718245 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:20.218233864 +0000 UTC m=+158.708296171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.718536 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0390146f-bbac-4efb-aaef-c7e9c41f8d51-catalog-content\") pod \"community-operators-7x8v8\" (UID: \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\") " pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.718679 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0390146f-bbac-4efb-aaef-c7e9c41f8d51-utilities\") pod \"community-operators-7x8v8\" (UID: \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\") " pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.818666 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.818848 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:20.318821888 +0000 UTC m=+158.808884205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.819014 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.819297 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:20.319289671 +0000 UTC m=+158.809351988 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.873744 4857 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 01 21:37:19 crc kubenswrapper[4857]: I1201 21:37:19.920090 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:19 crc kubenswrapper[4857]: E1201 21:37:19.920775 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 21:37:20.420742378 +0000 UTC m=+158.910804695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.094292 4857 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-01T21:37:19.873760912Z","Handler":null,"Name":""} Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.096034 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:20 crc kubenswrapper[4857]: E1201 21:37:20.108338 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 21:37:20.608322903 +0000 UTC m=+159.098385220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zdzg5" (UID: "658992cc-be88-49f4-b35c-bc445583fc39") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.108778 4857 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.108812 4857 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.167815 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6zrm\" (UniqueName: \"kubernetes.io/projected/0390146f-bbac-4efb-aaef-c7e9c41f8d51-kube-api-access-q6zrm\") pod \"community-operators-7x8v8\" (UID: \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\") " pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.217108 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.268418 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.319799 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.329072 4857 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.329111 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.534517 4857 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/community-operators-k9szd" secret="" err="failed to sync secret cache: timed out waiting for the condition" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.534594 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.536244 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:20 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:20 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:20 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.536279 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.551515 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6mll7" event={"ID":"5361e52d-63b2-4f66-b786-9b1348a6b428","Type":"ContainerStarted","Data":"b2cf92494934bc5b6a33ee495d575beea8fce9a3932f16f67d49eb79016d5ea4"} Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.665322 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.668102 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zdzg5\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.674234 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.877215 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.877257 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.887201 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-w54ht container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.887259 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-w54ht" podUID="0fd59b67-77d3-4e4f-b0b4-4bc714e29e73" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.887352 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-w54ht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.887406 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w54ht" podUID="0fd59b67-77d3-4e4f-b0b4-4bc714e29e73" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.903636 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4s5mw"] Dec 01 21:37:20 crc kubenswrapper[4857]: I1201 21:37:20.917337 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.011443 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.050103 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r2rmn"] Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.215207 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-thg27"] Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.226185 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-thg27"] Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.226603 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.243657 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.336008 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fvpq\" (UniqueName: \"kubernetes.io/projected/2235e98c-f4f9-40ae-8f47-788c336e5893-kube-api-access-9fvpq\") pod \"redhat-marketplace-thg27\" (UID: \"2235e98c-f4f9-40ae-8f47-788c336e5893\") " pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.380395 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2235e98c-f4f9-40ae-8f47-788c336e5893-catalog-content\") pod \"redhat-marketplace-thg27\" (UID: \"2235e98c-f4f9-40ae-8f47-788c336e5893\") " pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.380515 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2235e98c-f4f9-40ae-8f47-788c336e5893-utilities\") pod \"redhat-marketplace-thg27\" (UID: \"2235e98c-f4f9-40ae-8f47-788c336e5893\") " pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.389689 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.456556 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.469439 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:21 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:21 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:21 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.469491 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.480171 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xrmc2" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.482625 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2235e98c-f4f9-40ae-8f47-788c336e5893-catalog-content\") pod \"redhat-marketplace-thg27\" (UID: \"2235e98c-f4f9-40ae-8f47-788c336e5893\") " pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.482810 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2235e98c-f4f9-40ae-8f47-788c336e5893-utilities\") pod \"redhat-marketplace-thg27\" (UID: \"2235e98c-f4f9-40ae-8f47-788c336e5893\") " pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.482940 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fvpq\" (UniqueName: \"kubernetes.io/projected/2235e98c-f4f9-40ae-8f47-788c336e5893-kube-api-access-9fvpq\") pod \"redhat-marketplace-thg27\" (UID: \"2235e98c-f4f9-40ae-8f47-788c336e5893\") " pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.487702 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2235e98c-f4f9-40ae-8f47-788c336e5893-utilities\") pod \"redhat-marketplace-thg27\" (UID: \"2235e98c-f4f9-40ae-8f47-788c336e5893\") " pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.487926 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2235e98c-f4f9-40ae-8f47-788c336e5893-catalog-content\") pod \"redhat-marketplace-thg27\" (UID: \"2235e98c-f4f9-40ae-8f47-788c336e5893\") " pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.534122 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fvpq\" (UniqueName: \"kubernetes.io/projected/2235e98c-f4f9-40ae-8f47-788c336e5893-kube-api-access-9fvpq\") pod \"redhat-marketplace-thg27\" (UID: \"2235e98c-f4f9-40ae-8f47-788c336e5893\") " pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.539616 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ntgbq"] Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.554775 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.563260 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntgbq"] Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.566251 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2rmn" event={"ID":"41882b8a-22d5-4283-a5e1-8a400181d9de","Type":"ContainerStarted","Data":"e28a86b5fef854c914b3a8293c617c66ab1d6680a1e871ebaaf5836600386eef"} Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.611378 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6mll7" event={"ID":"5361e52d-63b2-4f66-b786-9b1348a6b428","Type":"ContainerStarted","Data":"eb7dc66127cb41ab1d8b2038e2e874003b10680d5581e5c2a4266062bd8d8dd0"} Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.612798 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4s5mw" event={"ID":"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a","Type":"ContainerStarted","Data":"7f24f22b6e81f45b5b03ed5d918ad908d387f243bb946e159a3e2abd9333cdb8"} Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.618648 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k9szd"] Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.635841 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.703725 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/902b6911-fd58-475c-adb0-22cafe4a2180-catalog-content\") pod \"redhat-marketplace-ntgbq\" (UID: \"902b6911-fd58-475c-adb0-22cafe4a2180\") " pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.703828 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p7nl\" (UniqueName: \"kubernetes.io/projected/902b6911-fd58-475c-adb0-22cafe4a2180-kube-api-access-4p7nl\") pod \"redhat-marketplace-ntgbq\" (UID: \"902b6911-fd58-475c-adb0-22cafe4a2180\") " pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.703856 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/902b6911-fd58-475c-adb0-22cafe4a2180-utilities\") pod \"redhat-marketplace-ntgbq\" (UID: \"902b6911-fd58-475c-adb0-22cafe4a2180\") " pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.710657 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-6mll7" podStartSLOduration=13.710638617 podStartE2EDuration="13.710638617s" podCreationTimestamp="2025-12-01 21:37:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:21.708334584 +0000 UTC m=+160.198396901" watchObservedRunningTime="2025-12-01 21:37:21.710638617 +0000 UTC m=+160.200700934" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.807181 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p7nl\" (UniqueName: \"kubernetes.io/projected/902b6911-fd58-475c-adb0-22cafe4a2180-kube-api-access-4p7nl\") pod \"redhat-marketplace-ntgbq\" (UID: \"902b6911-fd58-475c-adb0-22cafe4a2180\") " pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.807638 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/902b6911-fd58-475c-adb0-22cafe4a2180-utilities\") pod \"redhat-marketplace-ntgbq\" (UID: \"902b6911-fd58-475c-adb0-22cafe4a2180\") " pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.807694 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/902b6911-fd58-475c-adb0-22cafe4a2180-catalog-content\") pod \"redhat-marketplace-ntgbq\" (UID: \"902b6911-fd58-475c-adb0-22cafe4a2180\") " pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.808296 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/902b6911-fd58-475c-adb0-22cafe4a2180-catalog-content\") pod \"redhat-marketplace-ntgbq\" (UID: \"902b6911-fd58-475c-adb0-22cafe4a2180\") " pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.808583 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/902b6911-fd58-475c-adb0-22cafe4a2180-utilities\") pod \"redhat-marketplace-ntgbq\" (UID: \"902b6911-fd58-475c-adb0-22cafe4a2180\") " pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.924739 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p7nl\" (UniqueName: \"kubernetes.io/projected/902b6911-fd58-475c-adb0-22cafe4a2180-kube-api-access-4p7nl\") pod \"redhat-marketplace-ntgbq\" (UID: \"902b6911-fd58-475c-adb0-22cafe4a2180\") " pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.934434 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 01 21:37:21 crc kubenswrapper[4857]: I1201 21:37:21.935863 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.088587 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7x8v8"] Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.157101 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zrwmc"] Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.158313 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.162486 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.225256 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zrwmc"] Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.293110 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zdzg5"] Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.325980 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd2bq\" (UniqueName: \"kubernetes.io/projected/226b6f72-5676-4dc8-9536-d99da01652c8-kube-api-access-dd2bq\") pod \"redhat-operators-zrwmc\" (UID: \"226b6f72-5676-4dc8-9536-d99da01652c8\") " pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.326057 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/226b6f72-5676-4dc8-9536-d99da01652c8-catalog-content\") pod \"redhat-operators-zrwmc\" (UID: \"226b6f72-5676-4dc8-9536-d99da01652c8\") " pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.326118 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/226b6f72-5676-4dc8-9536-d99da01652c8-utilities\") pod \"redhat-operators-zrwmc\" (UID: \"226b6f72-5676-4dc8-9536-d99da01652c8\") " pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.435219 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd2bq\" (UniqueName: \"kubernetes.io/projected/226b6f72-5676-4dc8-9536-d99da01652c8-kube-api-access-dd2bq\") pod \"redhat-operators-zrwmc\" (UID: \"226b6f72-5676-4dc8-9536-d99da01652c8\") " pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.435275 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/226b6f72-5676-4dc8-9536-d99da01652c8-catalog-content\") pod \"redhat-operators-zrwmc\" (UID: \"226b6f72-5676-4dc8-9536-d99da01652c8\") " pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.435337 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/226b6f72-5676-4dc8-9536-d99da01652c8-utilities\") pod \"redhat-operators-zrwmc\" (UID: \"226b6f72-5676-4dc8-9536-d99da01652c8\") " pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.436017 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/226b6f72-5676-4dc8-9536-d99da01652c8-utilities\") pod \"redhat-operators-zrwmc\" (UID: \"226b6f72-5676-4dc8-9536-d99da01652c8\") " pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.436698 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/226b6f72-5676-4dc8-9536-d99da01652c8-catalog-content\") pod \"redhat-operators-zrwmc\" (UID: \"226b6f72-5676-4dc8-9536-d99da01652c8\") " pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.449186 4857 patch_prober.go:28] interesting pod/apiserver-76f77b778f-g6p59 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 01 21:37:22 crc kubenswrapper[4857]: [+]log ok Dec 01 21:37:22 crc kubenswrapper[4857]: [+]etcd ok Dec 01 21:37:22 crc kubenswrapper[4857]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 01 21:37:22 crc kubenswrapper[4857]: [+]poststarthook/generic-apiserver-start-informers ok Dec 01 21:37:22 crc kubenswrapper[4857]: [+]poststarthook/max-in-flight-filter ok Dec 01 21:37:22 crc kubenswrapper[4857]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 01 21:37:22 crc kubenswrapper[4857]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 01 21:37:22 crc kubenswrapper[4857]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 01 21:37:22 crc kubenswrapper[4857]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 01 21:37:22 crc kubenswrapper[4857]: [+]poststarthook/project.openshift.io-projectcache ok Dec 01 21:37:22 crc kubenswrapper[4857]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 01 21:37:22 crc kubenswrapper[4857]: [+]poststarthook/openshift.io-startinformers ok Dec 01 21:37:22 crc kubenswrapper[4857]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 01 21:37:22 crc kubenswrapper[4857]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 01 21:37:22 crc kubenswrapper[4857]: livez check failed Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.449283 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-g6p59" podUID="b6911089-a8c9-4e07-8f85-5d5d2ff9f736" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.471254 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:22 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:22 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:22 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.471308 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.487901 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd2bq\" (UniqueName: \"kubernetes.io/projected/226b6f72-5676-4dc8-9536-d99da01652c8-kube-api-access-dd2bq\") pod \"redhat-operators-zrwmc\" (UID: \"226b6f72-5676-4dc8-9536-d99da01652c8\") " pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.493181 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.493645 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.498512 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.501523 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.519490 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.526509 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.550724 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-htcng"] Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.551738 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.620418 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-htcng"] Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.654591 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7q2n\" (UniqueName: \"kubernetes.io/projected/a4e9e374-0931-4274-8bc4-8a9c5e405578-kube-api-access-q7q2n\") pod \"redhat-operators-htcng\" (UID: \"a4e9e374-0931-4274-8bc4-8a9c5e405578\") " pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.654647 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4e9e374-0931-4274-8bc4-8a9c5e405578-utilities\") pod \"redhat-operators-htcng\" (UID: \"a4e9e374-0931-4274-8bc4-8a9c5e405578\") " pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.654668 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4e9e374-0931-4274-8bc4-8a9c5e405578-catalog-content\") pod \"redhat-operators-htcng\" (UID: \"a4e9e374-0931-4274-8bc4-8a9c5e405578\") " pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.654710 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31a1c801-c108-45bf-8841-5a0cf11467c9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"31a1c801-c108-45bf-8841-5a0cf11467c9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.654769 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31a1c801-c108-45bf-8841-5a0cf11467c9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"31a1c801-c108-45bf-8841-5a0cf11467c9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.699177 4857 generic.go:334] "Generic (PLEG): container finished" podID="8049f37f-5448-4796-b384-691649bd559c" containerID="416d990ea4600370647de5c70a400a0db0f4bcc98d93519e4132ed021ef0f2e4" exitCode=0 Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.699274 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9szd" event={"ID":"8049f37f-5448-4796-b384-691649bd559c","Type":"ContainerDied","Data":"416d990ea4600370647de5c70a400a0db0f4bcc98d93519e4132ed021ef0f2e4"} Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.699300 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9szd" event={"ID":"8049f37f-5448-4796-b384-691649bd559c","Type":"ContainerStarted","Data":"67667efc4b7857e24fb75b7f672ccce6c0167a8ec2385a0f5e1e19546291f49f"} Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.714247 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.718205 4857 generic.go:334] "Generic (PLEG): container finished" podID="41882b8a-22d5-4283-a5e1-8a400181d9de" containerID="5cfc2f8fa4a7d4883db82d8dd92c89440dee18dd62d7bdc44314387953cb0717" exitCode=0 Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.718610 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2rmn" event={"ID":"41882b8a-22d5-4283-a5e1-8a400181d9de","Type":"ContainerDied","Data":"5cfc2f8fa4a7d4883db82d8dd92c89440dee18dd62d7bdc44314387953cb0717"} Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.740319 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-thg27"] Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.757279 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31a1c801-c108-45bf-8841-5a0cf11467c9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"31a1c801-c108-45bf-8841-5a0cf11467c9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.757339 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7q2n\" (UniqueName: \"kubernetes.io/projected/a4e9e374-0931-4274-8bc4-8a9c5e405578-kube-api-access-q7q2n\") pod \"redhat-operators-htcng\" (UID: \"a4e9e374-0931-4274-8bc4-8a9c5e405578\") " pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.757375 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4e9e374-0931-4274-8bc4-8a9c5e405578-utilities\") pod \"redhat-operators-htcng\" (UID: \"a4e9e374-0931-4274-8bc4-8a9c5e405578\") " pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.757396 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4e9e374-0931-4274-8bc4-8a9c5e405578-catalog-content\") pod \"redhat-operators-htcng\" (UID: \"a4e9e374-0931-4274-8bc4-8a9c5e405578\") " pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.757432 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31a1c801-c108-45bf-8841-5a0cf11467c9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"31a1c801-c108-45bf-8841-5a0cf11467c9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.758273 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31a1c801-c108-45bf-8841-5a0cf11467c9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"31a1c801-c108-45bf-8841-5a0cf11467c9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.758798 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4e9e374-0931-4274-8bc4-8a9c5e405578-utilities\") pod \"redhat-operators-htcng\" (UID: \"a4e9e374-0931-4274-8bc4-8a9c5e405578\") " pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.764706 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4e9e374-0931-4274-8bc4-8a9c5e405578-catalog-content\") pod \"redhat-operators-htcng\" (UID: \"a4e9e374-0931-4274-8bc4-8a9c5e405578\") " pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.765539 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" event={"ID":"658992cc-be88-49f4-b35c-bc445583fc39","Type":"ContainerStarted","Data":"2b469c820b65d666f421b57cbc052f8336afa928a6a52f2aeeba9f525091e944"} Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.786766 4857 generic.go:334] "Generic (PLEG): container finished" podID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" containerID="58e1feea57545e597fcbc11f62041bd0f69a2f9188d4f5d9404716bde32e7c17" exitCode=0 Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.786872 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4s5mw" event={"ID":"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a","Type":"ContainerDied","Data":"58e1feea57545e597fcbc11f62041bd0f69a2f9188d4f5d9404716bde32e7c17"} Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.804377 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7q2n\" (UniqueName: \"kubernetes.io/projected/a4e9e374-0931-4274-8bc4-8a9c5e405578-kube-api-access-q7q2n\") pod \"redhat-operators-htcng\" (UID: \"a4e9e374-0931-4274-8bc4-8a9c5e405578\") " pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.806866 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31a1c801-c108-45bf-8841-5a0cf11467c9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"31a1c801-c108-45bf-8841-5a0cf11467c9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.810971 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntgbq"] Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.845597 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7x8v8" event={"ID":"0390146f-bbac-4efb-aaef-c7e9c41f8d51","Type":"ContainerStarted","Data":"70e10783aba9cd279cb92faab8b281e264ad9939ad04b08bb7a8976a1b6a1583"} Dec 01 21:37:22 crc kubenswrapper[4857]: I1201 21:37:22.894984 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.039904 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.148082 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zrwmc"] Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.506255 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:23 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:23 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:23 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.506619 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.523519 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 21:37:23 crc kubenswrapper[4857]: W1201 21:37:23.529450 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod31a1c801_c108_45bf_8841_5a0cf11467c9.slice/crio-fcfcaa9571920ac0d8151fe2ec5c17c9288a31f8e4c92e08c1254d7d9470af75 WatchSource:0}: Error finding container fcfcaa9571920ac0d8151fe2ec5c17c9288a31f8e4c92e08c1254d7d9470af75: Status 404 returned error can't find the container with id fcfcaa9571920ac0d8151fe2ec5c17c9288a31f8e4c92e08c1254d7d9470af75 Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.555709 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-htcng"] Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.876550 4857 generic.go:334] "Generic (PLEG): container finished" podID="2235e98c-f4f9-40ae-8f47-788c336e5893" containerID="76a33f14ae5a7f7cca7acbe239f2b8e6300df462ba9bc33a262fcc1c2912ad47" exitCode=0 Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.883608 4857 generic.go:334] "Generic (PLEG): container finished" podID="902b6911-fd58-475c-adb0-22cafe4a2180" containerID="51c8e3031385eb25a4a560b1474e0b40771a837a44229d576109baddfc72a411" exitCode=0 Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.902738 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.904854 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.904889 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thg27" event={"ID":"2235e98c-f4f9-40ae-8f47-788c336e5893","Type":"ContainerDied","Data":"76a33f14ae5a7f7cca7acbe239f2b8e6300df462ba9bc33a262fcc1c2912ad47"} Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.904910 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thg27" event={"ID":"2235e98c-f4f9-40ae-8f47-788c336e5893","Type":"ContainerStarted","Data":"b1b516c485e262175113e0f6e35038996f6b7453acff478d58a2fd4beefec640"} Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.904932 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"31a1c801-c108-45bf-8841-5a0cf11467c9","Type":"ContainerStarted","Data":"fcfcaa9571920ac0d8151fe2ec5c17c9288a31f8e4c92e08c1254d7d9470af75"} Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.904947 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntgbq" event={"ID":"902b6911-fd58-475c-adb0-22cafe4a2180","Type":"ContainerDied","Data":"51c8e3031385eb25a4a560b1474e0b40771a837a44229d576109baddfc72a411"} Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.904959 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntgbq" event={"ID":"902b6911-fd58-475c-adb0-22cafe4a2180","Type":"ContainerStarted","Data":"6aefb605f627d75803ce20f3ac344f97fbbb1a7fb6b5a0f0c9d7b40a510c6e8b"} Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.904972 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htcng" event={"ID":"a4e9e374-0931-4274-8bc4-8a9c5e405578","Type":"ContainerStarted","Data":"2f648edb9bccdeb186beccbaea3c5c76a51d43495e1af0628fa18a4458788e2b"} Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.904983 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" event={"ID":"658992cc-be88-49f4-b35c-bc445583fc39","Type":"ContainerStarted","Data":"28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e"} Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.911818 4857 generic.go:334] "Generic (PLEG): container finished" podID="226b6f72-5676-4dc8-9536-d99da01652c8" containerID="2f9577782abbc8097923870aa3f701119563192fc8c56de31690da853351ee3c" exitCode=0 Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.912763 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrwmc" event={"ID":"226b6f72-5676-4dc8-9536-d99da01652c8","Type":"ContainerDied","Data":"2f9577782abbc8097923870aa3f701119563192fc8c56de31690da853351ee3c"} Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.912788 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrwmc" event={"ID":"226b6f72-5676-4dc8-9536-d99da01652c8","Type":"ContainerStarted","Data":"8446b2a1eeb243d5baeb89e1ed393cc9469b8fc513c38444416a9a160fef3720"} Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.918318 4857 generic.go:334] "Generic (PLEG): container finished" podID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" containerID="2d83cbf2bc6f2d24c3943aaf996c9c9d0cd3810cace60b3f1f3ace8987ace44d" exitCode=0 Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.918362 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7x8v8" event={"ID":"0390146f-bbac-4efb-aaef-c7e9c41f8d51","Type":"ContainerDied","Data":"2d83cbf2bc6f2d24c3943aaf996c9c9d0cd3810cace60b3f1f3ace8987ace44d"} Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.929659 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/485d0d95-8002-4bc3-8935-8b8f4ab5b2b3-metrics-certs\") pod \"network-metrics-daemon-n2fsz\" (UID: \"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3\") " pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:37:23 crc kubenswrapper[4857]: I1201 21:37:23.964502 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n2fsz" Dec 01 21:37:24 crc kubenswrapper[4857]: I1201 21:37:24.060566 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" podStartSLOduration=144.060549297 podStartE2EDuration="2m24.060549297s" podCreationTimestamp="2025-12-01 21:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:23.94814608 +0000 UTC m=+162.438208397" watchObservedRunningTime="2025-12-01 21:37:24.060549297 +0000 UTC m=+162.550611614" Dec 01 21:37:24 crc kubenswrapper[4857]: I1201 21:37:24.531378 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:24 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:24 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:24 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:24 crc kubenswrapper[4857]: I1201 21:37:24.531502 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:24 crc kubenswrapper[4857]: I1201 21:37:24.687524 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-n2fsz"] Dec 01 21:37:24 crc kubenswrapper[4857]: W1201 21:37:24.733552 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod485d0d95_8002_4bc3_8935_8b8f4ab5b2b3.slice/crio-64a950cfc9e13e0183995ffe5668977a50c970b390de40727fbb1d9a3dfc6b3c WatchSource:0}: Error finding container 64a950cfc9e13e0183995ffe5668977a50c970b390de40727fbb1d9a3dfc6b3c: Status 404 returned error can't find the container with id 64a950cfc9e13e0183995ffe5668977a50c970b390de40727fbb1d9a3dfc6b3c Dec 01 21:37:24 crc kubenswrapper[4857]: I1201 21:37:24.938640 4857 generic.go:334] "Generic (PLEG): container finished" podID="a4e9e374-0931-4274-8bc4-8a9c5e405578" containerID="231308b248a8313c68b9b700ade180f803db673c39ee58616469d76955a0adcc" exitCode=0 Dec 01 21:37:24 crc kubenswrapper[4857]: I1201 21:37:24.938737 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htcng" event={"ID":"a4e9e374-0931-4274-8bc4-8a9c5e405578","Type":"ContainerDied","Data":"231308b248a8313c68b9b700ade180f803db673c39ee58616469d76955a0adcc"} Dec 01 21:37:24 crc kubenswrapper[4857]: I1201 21:37:24.948736 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" event={"ID":"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3","Type":"ContainerStarted","Data":"64a950cfc9e13e0183995ffe5668977a50c970b390de40727fbb1d9a3dfc6b3c"} Dec 01 21:37:24 crc kubenswrapper[4857]: I1201 21:37:24.952363 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"31a1c801-c108-45bf-8841-5a0cf11467c9","Type":"ContainerStarted","Data":"ec8e4bc63c22a3c9942418dce975137938fbbb894557b43b60bcbd0b968d19c9"} Dec 01 21:37:24 crc kubenswrapper[4857]: I1201 21:37:24.974927 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.974906917 podStartE2EDuration="2.974906917s" podCreationTimestamp="2025-12-01 21:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:24.970882377 +0000 UTC m=+163.460944714" watchObservedRunningTime="2025-12-01 21:37:24.974906917 +0000 UTC m=+163.464969234" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.460536 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:25 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:25 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:25 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.460590 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.654937 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.655780 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.674319 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.674432 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.674929 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.715758 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3949cdc4-c554-4c44-b234-715b7bd1ae73-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3949cdc4-c554-4c44-b234-715b7bd1ae73\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.715834 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3949cdc4-c554-4c44-b234-715b7bd1ae73-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3949cdc4-c554-4c44-b234-715b7bd1ae73\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.817773 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3949cdc4-c554-4c44-b234-715b7bd1ae73-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3949cdc4-c554-4c44-b234-715b7bd1ae73\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.818521 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3949cdc4-c554-4c44-b234-715b7bd1ae73-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3949cdc4-c554-4c44-b234-715b7bd1ae73\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.818641 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3949cdc4-c554-4c44-b234-715b7bd1ae73-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3949cdc4-c554-4c44-b234-715b7bd1ae73\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.841008 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3949cdc4-c554-4c44-b234-715b7bd1ae73-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3949cdc4-c554-4c44-b234-715b7bd1ae73\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.881409 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.890303 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-g6p59" Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.997771 4857 generic.go:334] "Generic (PLEG): container finished" podID="6c8375eb-1803-475e-b897-9307d7bc29ee" containerID="fdc6a1bc8c1714b9add47cbd440bbc801690cba6bde6e6b375970539e22fdff7" exitCode=0 Dec 01 21:37:25 crc kubenswrapper[4857]: I1201 21:37:25.998416 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" event={"ID":"6c8375eb-1803-475e-b897-9307d7bc29ee","Type":"ContainerDied","Data":"fdc6a1bc8c1714b9add47cbd440bbc801690cba6bde6e6b375970539e22fdff7"} Dec 01 21:37:26 crc kubenswrapper[4857]: I1201 21:37:26.084320 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 21:37:26 crc kubenswrapper[4857]: I1201 21:37:26.457600 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-tgwpg" Dec 01 21:37:26 crc kubenswrapper[4857]: I1201 21:37:26.459344 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:26 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:26 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:26 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:26 crc kubenswrapper[4857]: I1201 21:37:26.459379 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:26 crc kubenswrapper[4857]: I1201 21:37:26.959590 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 21:37:27 crc kubenswrapper[4857]: I1201 21:37:27.012608 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" event={"ID":"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3","Type":"ContainerStarted","Data":"71644aa8924d5fec44d4149f7f7e0fb4a66627a10bfa1fa118741b99ea6d7470"} Dec 01 21:37:27 crc kubenswrapper[4857]: I1201 21:37:27.014637 4857 generic.go:334] "Generic (PLEG): container finished" podID="31a1c801-c108-45bf-8841-5a0cf11467c9" containerID="ec8e4bc63c22a3c9942418dce975137938fbbb894557b43b60bcbd0b968d19c9" exitCode=0 Dec 01 21:37:27 crc kubenswrapper[4857]: I1201 21:37:27.014692 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"31a1c801-c108-45bf-8841-5a0cf11467c9","Type":"ContainerDied","Data":"ec8e4bc63c22a3c9942418dce975137938fbbb894557b43b60bcbd0b968d19c9"} Dec 01 21:37:27 crc kubenswrapper[4857]: I1201 21:37:27.016501 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3949cdc4-c554-4c44-b234-715b7bd1ae73","Type":"ContainerStarted","Data":"fd5c06a142dd7bbb4724ba40de5f90aa20f6d7f49a7448ed06ef90622fa720a9"} Dec 01 21:37:27 crc kubenswrapper[4857]: I1201 21:37:27.792853 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:27 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:27 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:27 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:27 crc kubenswrapper[4857]: I1201 21:37:27.793574 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:27 crc kubenswrapper[4857]: I1201 21:37:27.805954 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:37:27 crc kubenswrapper[4857]: I1201 21:37:27.806011 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.103363 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3949cdc4-c554-4c44-b234-715b7bd1ae73","Type":"ContainerStarted","Data":"5ad63a4f7aba75c389dfc14aa9f87f2785c542735573cb059d3d7be67528e371"} Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.126421 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n2fsz" event={"ID":"485d0d95-8002-4bc3-8935-8b8f4ab5b2b3","Type":"ContainerStarted","Data":"535d735071443722379e0aa49e2f3d4ea11737fb54311dbea006ade65b22b13d"} Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.246355 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.269759 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-n2fsz" podStartSLOduration=147.2696163 podStartE2EDuration="2m27.2696163s" podCreationTimestamp="2025-12-01 21:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:37:28.221657509 +0000 UTC m=+166.711719826" watchObservedRunningTime="2025-12-01 21:37:28.2696163 +0000 UTC m=+166.759678617" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.334074 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq6hw\" (UniqueName: \"kubernetes.io/projected/6c8375eb-1803-475e-b897-9307d7bc29ee-kube-api-access-nq6hw\") pod \"6c8375eb-1803-475e-b897-9307d7bc29ee\" (UID: \"6c8375eb-1803-475e-b897-9307d7bc29ee\") " Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.334144 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c8375eb-1803-475e-b897-9307d7bc29ee-config-volume\") pod \"6c8375eb-1803-475e-b897-9307d7bc29ee\" (UID: \"6c8375eb-1803-475e-b897-9307d7bc29ee\") " Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.334186 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c8375eb-1803-475e-b897-9307d7bc29ee-secret-volume\") pod \"6c8375eb-1803-475e-b897-9307d7bc29ee\" (UID: \"6c8375eb-1803-475e-b897-9307d7bc29ee\") " Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.342022 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c8375eb-1803-475e-b897-9307d7bc29ee-config-volume" (OuterVolumeSpecName: "config-volume") pod "6c8375eb-1803-475e-b897-9307d7bc29ee" (UID: "6c8375eb-1803-475e-b897-9307d7bc29ee"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.435486 4857 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c8375eb-1803-475e-b897-9307d7bc29ee-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.519296 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c8375eb-1803-475e-b897-9307d7bc29ee-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6c8375eb-1803-475e-b897-9307d7bc29ee" (UID: "6c8375eb-1803-475e-b897-9307d7bc29ee"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.519341 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c8375eb-1803-475e-b897-9307d7bc29ee-kube-api-access-nq6hw" (OuterVolumeSpecName: "kube-api-access-nq6hw") pod "6c8375eb-1803-475e-b897-9307d7bc29ee" (UID: "6c8375eb-1803-475e-b897-9307d7bc29ee"). InnerVolumeSpecName "kube-api-access-nq6hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.520506 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:28 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:28 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:28 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.520546 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.536471 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq6hw\" (UniqueName: \"kubernetes.io/projected/6c8375eb-1803-475e-b897-9307d7bc29ee-kube-api-access-nq6hw\") on node \"crc\" DevicePath \"\"" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.536515 4857 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c8375eb-1803-475e-b897-9307d7bc29ee-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.732279 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.839113 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31a1c801-c108-45bf-8841-5a0cf11467c9-kube-api-access\") pod \"31a1c801-c108-45bf-8841-5a0cf11467c9\" (UID: \"31a1c801-c108-45bf-8841-5a0cf11467c9\") " Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.839559 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31a1c801-c108-45bf-8841-5a0cf11467c9-kubelet-dir\") pod \"31a1c801-c108-45bf-8841-5a0cf11467c9\" (UID: \"31a1c801-c108-45bf-8841-5a0cf11467c9\") " Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.839709 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31a1c801-c108-45bf-8841-5a0cf11467c9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "31a1c801-c108-45bf-8841-5a0cf11467c9" (UID: "31a1c801-c108-45bf-8841-5a0cf11467c9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.840152 4857 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31a1c801-c108-45bf-8841-5a0cf11467c9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 21:37:28 crc kubenswrapper[4857]: I1201 21:37:28.944884 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31a1c801-c108-45bf-8841-5a0cf11467c9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "31a1c801-c108-45bf-8841-5a0cf11467c9" (UID: "31a1c801-c108-45bf-8841-5a0cf11467c9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.018886 4857 patch_prober.go:28] interesting pod/console-f9d7485db-r5kwb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.018939 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-r5kwb" podUID="b8e5475f-0369-448a-895e-c02b78884677" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.073071 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31a1c801-c108-45bf-8841-5a0cf11467c9-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.157743 4857 generic.go:334] "Generic (PLEG): container finished" podID="3949cdc4-c554-4c44-b234-715b7bd1ae73" containerID="5ad63a4f7aba75c389dfc14aa9f87f2785c542735573cb059d3d7be67528e371" exitCode=0 Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.157815 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3949cdc4-c554-4c44-b234-715b7bd1ae73","Type":"ContainerDied","Data":"5ad63a4f7aba75c389dfc14aa9f87f2785c542735573cb059d3d7be67528e371"} Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.162648 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"31a1c801-c108-45bf-8841-5a0cf11467c9","Type":"ContainerDied","Data":"fcfcaa9571920ac0d8151fe2ec5c17c9288a31f8e4c92e08c1254d7d9470af75"} Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.162687 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.162702 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcfcaa9571920ac0d8151fe2ec5c17c9288a31f8e4c92e08c1254d7d9470af75" Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.168620 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" event={"ID":"6c8375eb-1803-475e-b897-9307d7bc29ee","Type":"ContainerDied","Data":"8cab48f9ff506e8fff1cab5058b3294002b336b6eea39817dae02ac5a5cb37c7"} Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.168662 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb" Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.168680 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cab48f9ff506e8fff1cab5058b3294002b336b6eea39817dae02ac5a5cb37c7" Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.456570 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:29 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:29 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:29 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:29 crc kubenswrapper[4857]: I1201 21:37:29.456629 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:30 crc kubenswrapper[4857]: I1201 21:37:30.467657 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:30 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:30 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:30 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:30 crc kubenswrapper[4857]: I1201 21:37:30.467712 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:30 crc kubenswrapper[4857]: I1201 21:37:30.898690 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-w54ht container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 01 21:37:30 crc kubenswrapper[4857]: I1201 21:37:30.898990 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-w54ht" podUID="0fd59b67-77d3-4e4f-b0b4-4bc714e29e73" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 01 21:37:30 crc kubenswrapper[4857]: I1201 21:37:30.898765 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-w54ht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 01 21:37:30 crc kubenswrapper[4857]: I1201 21:37:30.899122 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w54ht" podUID="0fd59b67-77d3-4e4f-b0b4-4bc714e29e73" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 01 21:37:30 crc kubenswrapper[4857]: I1201 21:37:30.976406 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 21:37:31 crc kubenswrapper[4857]: I1201 21:37:31.100667 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3949cdc4-c554-4c44-b234-715b7bd1ae73-kube-api-access\") pod \"3949cdc4-c554-4c44-b234-715b7bd1ae73\" (UID: \"3949cdc4-c554-4c44-b234-715b7bd1ae73\") " Dec 01 21:37:31 crc kubenswrapper[4857]: I1201 21:37:31.100787 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3949cdc4-c554-4c44-b234-715b7bd1ae73-kubelet-dir\") pod \"3949cdc4-c554-4c44-b234-715b7bd1ae73\" (UID: \"3949cdc4-c554-4c44-b234-715b7bd1ae73\") " Dec 01 21:37:31 crc kubenswrapper[4857]: I1201 21:37:31.101174 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3949cdc4-c554-4c44-b234-715b7bd1ae73-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3949cdc4-c554-4c44-b234-715b7bd1ae73" (UID: "3949cdc4-c554-4c44-b234-715b7bd1ae73"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:37:31 crc kubenswrapper[4857]: I1201 21:37:31.131189 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3949cdc4-c554-4c44-b234-715b7bd1ae73-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3949cdc4-c554-4c44-b234-715b7bd1ae73" (UID: "3949cdc4-c554-4c44-b234-715b7bd1ae73"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:37:31 crc kubenswrapper[4857]: I1201 21:37:31.202798 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3949cdc4-c554-4c44-b234-715b7bd1ae73-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 21:37:31 crc kubenswrapper[4857]: I1201 21:37:31.202841 4857 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3949cdc4-c554-4c44-b234-715b7bd1ae73-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 21:37:31 crc kubenswrapper[4857]: I1201 21:37:31.287469 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3949cdc4-c554-4c44-b234-715b7bd1ae73","Type":"ContainerDied","Data":"fd5c06a142dd7bbb4724ba40de5f90aa20f6d7f49a7448ed06ef90622fa720a9"} Dec 01 21:37:31 crc kubenswrapper[4857]: I1201 21:37:31.287545 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd5c06a142dd7bbb4724ba40de5f90aa20f6d7f49a7448ed06ef90622fa720a9" Dec 01 21:37:31 crc kubenswrapper[4857]: I1201 21:37:31.287544 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 21:37:31 crc kubenswrapper[4857]: I1201 21:37:31.533086 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:31 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:31 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:31 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:31 crc kubenswrapper[4857]: I1201 21:37:31.533220 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:32 crc kubenswrapper[4857]: I1201 21:37:32.456858 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:32 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:32 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:32 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:32 crc kubenswrapper[4857]: I1201 21:37:32.457297 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:33 crc kubenswrapper[4857]: I1201 21:37:33.458116 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:33 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:33 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:33 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:33 crc kubenswrapper[4857]: I1201 21:37:33.458193 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:34 crc kubenswrapper[4857]: I1201 21:37:34.526486 4857 patch_prober.go:28] interesting pod/router-default-5444994796-z26dc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 21:37:34 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Dec 01 21:37:34 crc kubenswrapper[4857]: [+]process-running ok Dec 01 21:37:34 crc kubenswrapper[4857]: healthz check failed Dec 01 21:37:34 crc kubenswrapper[4857]: I1201 21:37:34.526544 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z26dc" podUID="555f0013-cb67-45f1-b96b-3f2b938fba11" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 21:37:35 crc kubenswrapper[4857]: I1201 21:37:35.461497 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:35 crc kubenswrapper[4857]: I1201 21:37:35.464827 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-z26dc" Dec 01 21:37:38 crc kubenswrapper[4857]: I1201 21:37:38.557977 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 21:37:39 crc kubenswrapper[4857]: I1201 21:37:39.007137 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:39 crc kubenswrapper[4857]: I1201 21:37:39.011001 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:37:40 crc kubenswrapper[4857]: I1201 21:37:40.893110 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-w54ht" Dec 01 21:37:40 crc kubenswrapper[4857]: I1201 21:37:40.925805 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:37:51 crc kubenswrapper[4857]: I1201 21:37:51.502331 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n5slz" Dec 01 21:37:57 crc kubenswrapper[4857]: I1201 21:37:57.863756 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:37:57 crc kubenswrapper[4857]: I1201 21:37:57.864095 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.473949 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 21:37:59 crc kubenswrapper[4857]: E1201 21:37:59.474188 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31a1c801-c108-45bf-8841-5a0cf11467c9" containerName="pruner" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.474204 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="31a1c801-c108-45bf-8841-5a0cf11467c9" containerName="pruner" Dec 01 21:37:59 crc kubenswrapper[4857]: E1201 21:37:59.474219 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3949cdc4-c554-4c44-b234-715b7bd1ae73" containerName="pruner" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.474225 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3949cdc4-c554-4c44-b234-715b7bd1ae73" containerName="pruner" Dec 01 21:37:59 crc kubenswrapper[4857]: E1201 21:37:59.474236 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c8375eb-1803-475e-b897-9307d7bc29ee" containerName="collect-profiles" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.474242 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c8375eb-1803-475e-b897-9307d7bc29ee" containerName="collect-profiles" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.475684 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="31a1c801-c108-45bf-8841-5a0cf11467c9" containerName="pruner" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.475705 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c8375eb-1803-475e-b897-9307d7bc29ee" containerName="collect-profiles" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.475712 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="3949cdc4-c554-4c44-b234-715b7bd1ae73" containerName="pruner" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.476194 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.476286 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.479702 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.479993 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.639484 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9765efe-fedf-4dc9-b28f-095fedff216e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c9765efe-fedf-4dc9-b28f-095fedff216e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.639536 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9765efe-fedf-4dc9-b28f-095fedff216e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c9765efe-fedf-4dc9-b28f-095fedff216e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.741418 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9765efe-fedf-4dc9-b28f-095fedff216e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c9765efe-fedf-4dc9-b28f-095fedff216e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.741784 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9765efe-fedf-4dc9-b28f-095fedff216e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c9765efe-fedf-4dc9-b28f-095fedff216e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.741677 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9765efe-fedf-4dc9-b28f-095fedff216e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c9765efe-fedf-4dc9-b28f-095fedff216e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.766559 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9765efe-fedf-4dc9-b28f-095fedff216e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c9765efe-fedf-4dc9-b28f-095fedff216e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 21:37:59 crc kubenswrapper[4857]: I1201 21:37:59.804019 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 21:38:01 crc kubenswrapper[4857]: E1201 21:38:01.036727 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 01 21:38:01 crc kubenswrapper[4857]: E1201 21:38:01.037146 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dd2bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-zrwmc_openshift-marketplace(226b6f72-5676-4dc8-9536-d99da01652c8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 21:38:01 crc kubenswrapper[4857]: E1201 21:38:01.040206 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-zrwmc" podUID="226b6f72-5676-4dc8-9536-d99da01652c8" Dec 01 21:38:01 crc kubenswrapper[4857]: E1201 21:38:01.383561 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 01 21:38:01 crc kubenswrapper[4857]: E1201 21:38:01.383861 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q7q2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-htcng_openshift-marketplace(a4e9e374-0931-4274-8bc4-8a9c5e405578): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 21:38:01 crc kubenswrapper[4857]: E1201 21:38:01.385967 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-htcng" podUID="a4e9e374-0931-4274-8bc4-8a9c5e405578" Dec 01 21:38:03 crc kubenswrapper[4857]: E1201 21:38:03.289945 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-htcng" podUID="a4e9e374-0931-4274-8bc4-8a9c5e405578" Dec 01 21:38:03 crc kubenswrapper[4857]: E1201 21:38:03.289984 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-zrwmc" podUID="226b6f72-5676-4dc8-9536-d99da01652c8" Dec 01 21:38:03 crc kubenswrapper[4857]: E1201 21:38:03.473724 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 01 21:38:03 crc kubenswrapper[4857]: E1201 21:38:03.473921 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-szw7m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-4s5mw_openshift-marketplace(2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 21:38:03 crc kubenswrapper[4857]: E1201 21:38:03.475139 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-4s5mw" podUID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.061583 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.064949 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.079614 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.235218 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c350e136-f264-40ea-9b0e-9211dea68129-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c350e136-f264-40ea-9b0e-9211dea68129\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.236273 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c350e136-f264-40ea-9b0e-9211dea68129-kube-api-access\") pod \"installer-9-crc\" (UID: \"c350e136-f264-40ea-9b0e-9211dea68129\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.236609 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c350e136-f264-40ea-9b0e-9211dea68129-var-lock\") pod \"installer-9-crc\" (UID: \"c350e136-f264-40ea-9b0e-9211dea68129\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.338169 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c350e136-f264-40ea-9b0e-9211dea68129-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c350e136-f264-40ea-9b0e-9211dea68129\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.338695 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c350e136-f264-40ea-9b0e-9211dea68129-kube-api-access\") pod \"installer-9-crc\" (UID: \"c350e136-f264-40ea-9b0e-9211dea68129\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.338881 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c350e136-f264-40ea-9b0e-9211dea68129-var-lock\") pod \"installer-9-crc\" (UID: \"c350e136-f264-40ea-9b0e-9211dea68129\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.339094 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c350e136-f264-40ea-9b0e-9211dea68129-var-lock\") pod \"installer-9-crc\" (UID: \"c350e136-f264-40ea-9b0e-9211dea68129\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.338347 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c350e136-f264-40ea-9b0e-9211dea68129-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c350e136-f264-40ea-9b0e-9211dea68129\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.374341 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c350e136-f264-40ea-9b0e-9211dea68129-kube-api-access\") pod \"installer-9-crc\" (UID: \"c350e136-f264-40ea-9b0e-9211dea68129\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:05 crc kubenswrapper[4857]: I1201 21:38:05.403337 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:07 crc kubenswrapper[4857]: E1201 21:38:07.803798 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-4s5mw" podUID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" Dec 01 21:38:07 crc kubenswrapper[4857]: E1201 21:38:07.938644 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 21:38:07 crc kubenswrapper[4857]: E1201 21:38:07.939303 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4p7nl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ntgbq_openshift-marketplace(902b6911-fd58-475c-adb0-22cafe4a2180): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 21:38:07 crc kubenswrapper[4857]: E1201 21:38:07.940713 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-ntgbq" podUID="902b6911-fd58-475c-adb0-22cafe4a2180" Dec 01 21:38:08 crc kubenswrapper[4857]: E1201 21:38:08.022436 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 01 21:38:08 crc kubenswrapper[4857]: E1201 21:38:08.022623 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2htwv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-r2rmn_openshift-marketplace(41882b8a-22d5-4283-a5e1-8a400181d9de): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 21:38:08 crc kubenswrapper[4857]: E1201 21:38:08.023826 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-r2rmn" podUID="41882b8a-22d5-4283-a5e1-8a400181d9de" Dec 01 21:38:08 crc kubenswrapper[4857]: E1201 21:38:08.527023 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 21:38:08 crc kubenswrapper[4857]: E1201 21:38:08.527267 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9fvpq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-thg27_openshift-marketplace(2235e98c-f4f9-40ae-8f47-788c336e5893): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 21:38:08 crc kubenswrapper[4857]: E1201 21:38:08.528506 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-thg27" podUID="2235e98c-f4f9-40ae-8f47-788c336e5893" Dec 01 21:38:11 crc kubenswrapper[4857]: E1201 21:38:11.747440 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-r2rmn" podUID="41882b8a-22d5-4283-a5e1-8a400181d9de" Dec 01 21:38:11 crc kubenswrapper[4857]: E1201 21:38:11.747485 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-thg27" podUID="2235e98c-f4f9-40ae-8f47-788c336e5893" Dec 01 21:38:11 crc kubenswrapper[4857]: E1201 21:38:11.747595 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ntgbq" podUID="902b6911-fd58-475c-adb0-22cafe4a2180" Dec 01 21:38:11 crc kubenswrapper[4857]: E1201 21:38:11.851321 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 21:38:11 crc kubenswrapper[4857]: E1201 21:38:11.851830 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-phc6k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-k9szd_openshift-marketplace(8049f37f-5448-4796-b384-691649bd559c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 21:38:11 crc kubenswrapper[4857]: E1201 21:38:11.856126 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-k9szd" podUID="8049f37f-5448-4796-b384-691649bd559c" Dec 01 21:38:11 crc kubenswrapper[4857]: E1201 21:38:11.865345 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 21:38:11 crc kubenswrapper[4857]: E1201 21:38:11.865530 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q6zrm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-7x8v8_openshift-marketplace(0390146f-bbac-4efb-aaef-c7e9c41f8d51): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 21:38:11 crc kubenswrapper[4857]: E1201 21:38:11.866763 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-7x8v8" podUID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" Dec 01 21:38:12 crc kubenswrapper[4857]: I1201 21:38:12.220799 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 21:38:12 crc kubenswrapper[4857]: W1201 21:38:12.235737 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc350e136_f264_40ea_9b0e_9211dea68129.slice/crio-3db2375a7f1ae95b8d53a8cb13468d5fabb6243dc2e7745834ae187db9a5d774 WatchSource:0}: Error finding container 3db2375a7f1ae95b8d53a8cb13468d5fabb6243dc2e7745834ae187db9a5d774: Status 404 returned error can't find the container with id 3db2375a7f1ae95b8d53a8cb13468d5fabb6243dc2e7745834ae187db9a5d774 Dec 01 21:38:12 crc kubenswrapper[4857]: I1201 21:38:12.283146 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 21:38:12 crc kubenswrapper[4857]: W1201 21:38:12.289439 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc9765efe_fedf_4dc9_b28f_095fedff216e.slice/crio-91bb5b9160221bdcf2eb35fd302db81e8beac81a4bafe1ebbc533852a7ec3974 WatchSource:0}: Error finding container 91bb5b9160221bdcf2eb35fd302db81e8beac81a4bafe1ebbc533852a7ec3974: Status 404 returned error can't find the container with id 91bb5b9160221bdcf2eb35fd302db81e8beac81a4bafe1ebbc533852a7ec3974 Dec 01 21:38:12 crc kubenswrapper[4857]: I1201 21:38:12.623070 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c9765efe-fedf-4dc9-b28f-095fedff216e","Type":"ContainerStarted","Data":"91bb5b9160221bdcf2eb35fd302db81e8beac81a4bafe1ebbc533852a7ec3974"} Dec 01 21:38:12 crc kubenswrapper[4857]: I1201 21:38:12.624844 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c350e136-f264-40ea-9b0e-9211dea68129","Type":"ContainerStarted","Data":"3db2375a7f1ae95b8d53a8cb13468d5fabb6243dc2e7745834ae187db9a5d774"} Dec 01 21:38:12 crc kubenswrapper[4857]: E1201 21:38:12.627004 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-k9szd" podUID="8049f37f-5448-4796-b384-691649bd559c" Dec 01 21:38:12 crc kubenswrapper[4857]: E1201 21:38:12.627515 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-7x8v8" podUID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" Dec 01 21:38:13 crc kubenswrapper[4857]: I1201 21:38:13.632248 4857 generic.go:334] "Generic (PLEG): container finished" podID="c9765efe-fedf-4dc9-b28f-095fedff216e" containerID="ef8390037586c9394d3a72d3612ab75fc0bb296b76eae2acd6ab107b11bcee46" exitCode=0 Dec 01 21:38:13 crc kubenswrapper[4857]: I1201 21:38:13.632795 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c9765efe-fedf-4dc9-b28f-095fedff216e","Type":"ContainerDied","Data":"ef8390037586c9394d3a72d3612ab75fc0bb296b76eae2acd6ab107b11bcee46"} Dec 01 21:38:13 crc kubenswrapper[4857]: I1201 21:38:13.636534 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c350e136-f264-40ea-9b0e-9211dea68129","Type":"ContainerStarted","Data":"d943040e60619af528fafac613ed9807b10f7f1145b385ab05950957ddb1d225"} Dec 01 21:38:13 crc kubenswrapper[4857]: I1201 21:38:13.674425 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=8.67440027 podStartE2EDuration="8.67440027s" podCreationTimestamp="2025-12-01 21:38:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:38:13.671948776 +0000 UTC m=+212.162011103" watchObservedRunningTime="2025-12-01 21:38:13.67440027 +0000 UTC m=+212.164462597" Dec 01 21:38:14 crc kubenswrapper[4857]: I1201 21:38:14.918607 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 21:38:15 crc kubenswrapper[4857]: I1201 21:38:15.096856 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9765efe-fedf-4dc9-b28f-095fedff216e-kubelet-dir\") pod \"c9765efe-fedf-4dc9-b28f-095fedff216e\" (UID: \"c9765efe-fedf-4dc9-b28f-095fedff216e\") " Dec 01 21:38:15 crc kubenswrapper[4857]: I1201 21:38:15.097422 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9765efe-fedf-4dc9-b28f-095fedff216e-kube-api-access\") pod \"c9765efe-fedf-4dc9-b28f-095fedff216e\" (UID: \"c9765efe-fedf-4dc9-b28f-095fedff216e\") " Dec 01 21:38:15 crc kubenswrapper[4857]: I1201 21:38:15.097134 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c9765efe-fedf-4dc9-b28f-095fedff216e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c9765efe-fedf-4dc9-b28f-095fedff216e" (UID: "c9765efe-fedf-4dc9-b28f-095fedff216e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:38:15 crc kubenswrapper[4857]: I1201 21:38:15.105963 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9765efe-fedf-4dc9-b28f-095fedff216e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c9765efe-fedf-4dc9-b28f-095fedff216e" (UID: "c9765efe-fedf-4dc9-b28f-095fedff216e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:38:15 crc kubenswrapper[4857]: I1201 21:38:15.338481 4857 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9765efe-fedf-4dc9-b28f-095fedff216e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:15 crc kubenswrapper[4857]: I1201 21:38:15.338516 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9765efe-fedf-4dc9-b28f-095fedff216e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:15 crc kubenswrapper[4857]: I1201 21:38:15.649509 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 21:38:15 crc kubenswrapper[4857]: I1201 21:38:15.649455 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c9765efe-fedf-4dc9-b28f-095fedff216e","Type":"ContainerDied","Data":"91bb5b9160221bdcf2eb35fd302db81e8beac81a4bafe1ebbc533852a7ec3974"} Dec 01 21:38:15 crc kubenswrapper[4857]: I1201 21:38:15.650180 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91bb5b9160221bdcf2eb35fd302db81e8beac81a4bafe1ebbc533852a7ec3974" Dec 01 21:38:15 crc kubenswrapper[4857]: I1201 21:38:15.651589 4857 generic.go:334] "Generic (PLEG): container finished" podID="a4e9e374-0931-4274-8bc4-8a9c5e405578" containerID="702b3dfce1029d0cc9dfff6bec1699332ab8d473f059c2615831849c391540ad" exitCode=0 Dec 01 21:38:15 crc kubenswrapper[4857]: I1201 21:38:15.651641 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htcng" event={"ID":"a4e9e374-0931-4274-8bc4-8a9c5e405578","Type":"ContainerDied","Data":"702b3dfce1029d0cc9dfff6bec1699332ab8d473f059c2615831849c391540ad"} Dec 01 21:38:16 crc kubenswrapper[4857]: I1201 21:38:16.659749 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htcng" event={"ID":"a4e9e374-0931-4274-8bc4-8a9c5e405578","Type":"ContainerStarted","Data":"926a7e2afe166509a818baa920e69a432512fbdb2ba55ce6727fc42bb525e15b"} Dec 01 21:38:16 crc kubenswrapper[4857]: I1201 21:38:16.662208 4857 generic.go:334] "Generic (PLEG): container finished" podID="226b6f72-5676-4dc8-9536-d99da01652c8" containerID="ba8eeae978be367d4518757a0fee5ecc88932ebd5f3bcc7f17aaafae99638ef9" exitCode=0 Dec 01 21:38:16 crc kubenswrapper[4857]: I1201 21:38:16.662232 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrwmc" event={"ID":"226b6f72-5676-4dc8-9536-d99da01652c8","Type":"ContainerDied","Data":"ba8eeae978be367d4518757a0fee5ecc88932ebd5f3bcc7f17aaafae99638ef9"} Dec 01 21:38:16 crc kubenswrapper[4857]: I1201 21:38:16.715467 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-htcng" podStartSLOduration=3.5208142860000002 podStartE2EDuration="54.715438284s" podCreationTimestamp="2025-12-01 21:37:22 +0000 UTC" firstStartedPulling="2025-12-01 21:37:24.94214646 +0000 UTC m=+163.432208777" lastFinishedPulling="2025-12-01 21:38:16.136770448 +0000 UTC m=+214.626832775" observedRunningTime="2025-12-01 21:38:16.688291839 +0000 UTC m=+215.178354166" watchObservedRunningTime="2025-12-01 21:38:16.715438284 +0000 UTC m=+215.205500611" Dec 01 21:38:17 crc kubenswrapper[4857]: I1201 21:38:17.673632 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrwmc" event={"ID":"226b6f72-5676-4dc8-9536-d99da01652c8","Type":"ContainerStarted","Data":"f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf"} Dec 01 21:38:17 crc kubenswrapper[4857]: I1201 21:38:17.704196 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zrwmc" podStartSLOduration=2.3641345400000002 podStartE2EDuration="55.704170903s" podCreationTimestamp="2025-12-01 21:37:22 +0000 UTC" firstStartedPulling="2025-12-01 21:37:23.918171089 +0000 UTC m=+162.408233406" lastFinishedPulling="2025-12-01 21:38:17.258207422 +0000 UTC m=+215.748269769" observedRunningTime="2025-12-01 21:38:17.700174423 +0000 UTC m=+216.190236780" watchObservedRunningTime="2025-12-01 21:38:17.704170903 +0000 UTC m=+216.194233220" Dec 01 21:38:20 crc kubenswrapper[4857]: I1201 21:38:20.537025 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mmb9s"] Dec 01 21:38:22 crc kubenswrapper[4857]: I1201 21:38:22.494886 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:38:22 crc kubenswrapper[4857]: I1201 21:38:22.496474 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:38:23 crc kubenswrapper[4857]: I1201 21:38:23.040391 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:38:23 crc kubenswrapper[4857]: I1201 21:38:23.040438 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:38:23 crc kubenswrapper[4857]: I1201 21:38:23.094198 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:38:23 crc kubenswrapper[4857]: I1201 21:38:23.577770 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zrwmc" podUID="226b6f72-5676-4dc8-9536-d99da01652c8" containerName="registry-server" probeResult="failure" output=< Dec 01 21:38:23 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Dec 01 21:38:23 crc kubenswrapper[4857]: > Dec 01 21:38:23 crc kubenswrapper[4857]: I1201 21:38:23.743373 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:38:24 crc kubenswrapper[4857]: I1201 21:38:24.713140 4857 generic.go:334] "Generic (PLEG): container finished" podID="902b6911-fd58-475c-adb0-22cafe4a2180" containerID="86d0c1963a159089023764b3ffe77fb660bd9ec52c95613c717daf40a01f943b" exitCode=0 Dec 01 21:38:24 crc kubenswrapper[4857]: I1201 21:38:24.713248 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntgbq" event={"ID":"902b6911-fd58-475c-adb0-22cafe4a2180","Type":"ContainerDied","Data":"86d0c1963a159089023764b3ffe77fb660bd9ec52c95613c717daf40a01f943b"} Dec 01 21:38:24 crc kubenswrapper[4857]: I1201 21:38:24.720718 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4s5mw" event={"ID":"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a","Type":"ContainerStarted","Data":"bfbfc9af4f9f8fa180d661e2bab33cb070216b92e544c8487a17baee12962879"} Dec 01 21:38:25 crc kubenswrapper[4857]: I1201 21:38:25.737622 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7x8v8" event={"ID":"0390146f-bbac-4efb-aaef-c7e9c41f8d51","Type":"ContainerStarted","Data":"daa9a03301e1e58c40db0e87f032767563ebe48b57e46da1dd0754fc91eb7329"} Dec 01 21:38:25 crc kubenswrapper[4857]: I1201 21:38:25.740003 4857 generic.go:334] "Generic (PLEG): container finished" podID="8049f37f-5448-4796-b384-691649bd559c" containerID="4330ab6c16a16ed62e8ea9df65b84cbf1e29ab7f3560c278f9884f8ab04914a2" exitCode=0 Dec 01 21:38:25 crc kubenswrapper[4857]: I1201 21:38:25.740086 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9szd" event={"ID":"8049f37f-5448-4796-b384-691649bd559c","Type":"ContainerDied","Data":"4330ab6c16a16ed62e8ea9df65b84cbf1e29ab7f3560c278f9884f8ab04914a2"} Dec 01 21:38:25 crc kubenswrapper[4857]: I1201 21:38:25.745095 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntgbq" event={"ID":"902b6911-fd58-475c-adb0-22cafe4a2180","Type":"ContainerStarted","Data":"36407e50935ef0762111442c54972adb0e91618d5672628075721c7a008d5c6d"} Dec 01 21:38:25 crc kubenswrapper[4857]: I1201 21:38:25.747730 4857 generic.go:334] "Generic (PLEG): container finished" podID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" containerID="bfbfc9af4f9f8fa180d661e2bab33cb070216b92e544c8487a17baee12962879" exitCode=0 Dec 01 21:38:25 crc kubenswrapper[4857]: I1201 21:38:25.747775 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4s5mw" event={"ID":"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a","Type":"ContainerDied","Data":"bfbfc9af4f9f8fa180d661e2bab33cb070216b92e544c8487a17baee12962879"} Dec 01 21:38:25 crc kubenswrapper[4857]: I1201 21:38:25.808754 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ntgbq" podStartSLOduration=3.528860235 podStartE2EDuration="1m4.80872122s" podCreationTimestamp="2025-12-01 21:37:21 +0000 UTC" firstStartedPulling="2025-12-01 21:37:23.888210449 +0000 UTC m=+162.378272766" lastFinishedPulling="2025-12-01 21:38:25.168071434 +0000 UTC m=+223.658133751" observedRunningTime="2025-12-01 21:38:25.786859884 +0000 UTC m=+224.276922201" watchObservedRunningTime="2025-12-01 21:38:25.80872122 +0000 UTC m=+224.298783537" Dec 01 21:38:26 crc kubenswrapper[4857]: I1201 21:38:26.653534 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-htcng"] Dec 01 21:38:26 crc kubenswrapper[4857]: I1201 21:38:26.654193 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-htcng" podUID="a4e9e374-0931-4274-8bc4-8a9c5e405578" containerName="registry-server" containerID="cri-o://926a7e2afe166509a818baa920e69a432512fbdb2ba55ce6727fc42bb525e15b" gracePeriod=2 Dec 01 21:38:26 crc kubenswrapper[4857]: I1201 21:38:26.754167 4857 generic.go:334] "Generic (PLEG): container finished" podID="2235e98c-f4f9-40ae-8f47-788c336e5893" containerID="acdb4cdaf26d61dc3fd6b5cf99496a03cce2574a9542a00a1f4caf3858e2702b" exitCode=0 Dec 01 21:38:26 crc kubenswrapper[4857]: I1201 21:38:26.754242 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thg27" event={"ID":"2235e98c-f4f9-40ae-8f47-788c336e5893","Type":"ContainerDied","Data":"acdb4cdaf26d61dc3fd6b5cf99496a03cce2574a9542a00a1f4caf3858e2702b"} Dec 01 21:38:26 crc kubenswrapper[4857]: I1201 21:38:26.756573 4857 generic.go:334] "Generic (PLEG): container finished" podID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" containerID="daa9a03301e1e58c40db0e87f032767563ebe48b57e46da1dd0754fc91eb7329" exitCode=0 Dec 01 21:38:26 crc kubenswrapper[4857]: I1201 21:38:26.756614 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7x8v8" event={"ID":"0390146f-bbac-4efb-aaef-c7e9c41f8d51","Type":"ContainerDied","Data":"daa9a03301e1e58c40db0e87f032767563ebe48b57e46da1dd0754fc91eb7329"} Dec 01 21:38:26 crc kubenswrapper[4857]: I1201 21:38:26.759160 4857 generic.go:334] "Generic (PLEG): container finished" podID="41882b8a-22d5-4283-a5e1-8a400181d9de" containerID="2762ba9da4f468a30fffc2c8a642df5491bba0db3c30e40aa733ddeae717c6dd" exitCode=0 Dec 01 21:38:26 crc kubenswrapper[4857]: I1201 21:38:26.759186 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2rmn" event={"ID":"41882b8a-22d5-4283-a5e1-8a400181d9de","Type":"ContainerDied","Data":"2762ba9da4f468a30fffc2c8a642df5491bba0db3c30e40aa733ddeae717c6dd"} Dec 01 21:38:27 crc kubenswrapper[4857]: I1201 21:38:27.807435 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:38:27 crc kubenswrapper[4857]: I1201 21:38:27.807508 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:38:27 crc kubenswrapper[4857]: I1201 21:38:27.807561 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:38:27 crc kubenswrapper[4857]: I1201 21:38:27.808248 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 21:38:27 crc kubenswrapper[4857]: I1201 21:38:27.808369 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786" gracePeriod=600 Dec 01 21:38:28 crc kubenswrapper[4857]: I1201 21:38:28.772360 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786" exitCode=0 Dec 01 21:38:28 crc kubenswrapper[4857]: I1201 21:38:28.772444 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786"} Dec 01 21:38:28 crc kubenswrapper[4857]: I1201 21:38:28.775641 4857 generic.go:334] "Generic (PLEG): container finished" podID="a4e9e374-0931-4274-8bc4-8a9c5e405578" containerID="926a7e2afe166509a818baa920e69a432512fbdb2ba55ce6727fc42bb525e15b" exitCode=0 Dec 01 21:38:28 crc kubenswrapper[4857]: I1201 21:38:28.775673 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htcng" event={"ID":"a4e9e374-0931-4274-8bc4-8a9c5e405578","Type":"ContainerDied","Data":"926a7e2afe166509a818baa920e69a432512fbdb2ba55ce6727fc42bb525e15b"} Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.291658 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.408660 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4e9e374-0931-4274-8bc4-8a9c5e405578-utilities\") pod \"a4e9e374-0931-4274-8bc4-8a9c5e405578\" (UID: \"a4e9e374-0931-4274-8bc4-8a9c5e405578\") " Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.408715 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4e9e374-0931-4274-8bc4-8a9c5e405578-catalog-content\") pod \"a4e9e374-0931-4274-8bc4-8a9c5e405578\" (UID: \"a4e9e374-0931-4274-8bc4-8a9c5e405578\") " Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.408828 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7q2n\" (UniqueName: \"kubernetes.io/projected/a4e9e374-0931-4274-8bc4-8a9c5e405578-kube-api-access-q7q2n\") pod \"a4e9e374-0931-4274-8bc4-8a9c5e405578\" (UID: \"a4e9e374-0931-4274-8bc4-8a9c5e405578\") " Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.409627 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4e9e374-0931-4274-8bc4-8a9c5e405578-utilities" (OuterVolumeSpecName: "utilities") pod "a4e9e374-0931-4274-8bc4-8a9c5e405578" (UID: "a4e9e374-0931-4274-8bc4-8a9c5e405578"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.413472 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4e9e374-0931-4274-8bc4-8a9c5e405578-kube-api-access-q7q2n" (OuterVolumeSpecName: "kube-api-access-q7q2n") pod "a4e9e374-0931-4274-8bc4-8a9c5e405578" (UID: "a4e9e374-0931-4274-8bc4-8a9c5e405578"). InnerVolumeSpecName "kube-api-access-q7q2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.510648 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4e9e374-0931-4274-8bc4-8a9c5e405578-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.510673 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7q2n\" (UniqueName: \"kubernetes.io/projected/a4e9e374-0931-4274-8bc4-8a9c5e405578-kube-api-access-q7q2n\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.522782 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4e9e374-0931-4274-8bc4-8a9c5e405578-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4e9e374-0931-4274-8bc4-8a9c5e405578" (UID: "a4e9e374-0931-4274-8bc4-8a9c5e405578"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.611992 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4e9e374-0931-4274-8bc4-8a9c5e405578-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.787660 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"91968eb84103df465c19e9a8db52e323c98787b9caff59512702928cd30a7bb1"} Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.791417 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htcng" event={"ID":"a4e9e374-0931-4274-8bc4-8a9c5e405578","Type":"ContainerDied","Data":"2f648edb9bccdeb186beccbaea3c5c76a51d43495e1af0628fa18a4458788e2b"} Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.791474 4857 scope.go:117] "RemoveContainer" containerID="926a7e2afe166509a818baa920e69a432512fbdb2ba55ce6727fc42bb525e15b" Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.791546 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htcng" Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.874676 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-htcng"] Dec 01 21:38:29 crc kubenswrapper[4857]: I1201 21:38:29.877812 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-htcng"] Dec 01 21:38:30 crc kubenswrapper[4857]: I1201 21:38:30.508024 4857 scope.go:117] "RemoveContainer" containerID="702b3dfce1029d0cc9dfff6bec1699332ab8d473f059c2615831849c391540ad" Dec 01 21:38:30 crc kubenswrapper[4857]: I1201 21:38:30.800515 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4s5mw" event={"ID":"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a","Type":"ContainerStarted","Data":"6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f"} Dec 01 21:38:30 crc kubenswrapper[4857]: I1201 21:38:30.803208 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9szd" event={"ID":"8049f37f-5448-4796-b384-691649bd559c","Type":"ContainerStarted","Data":"511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5"} Dec 01 21:38:30 crc kubenswrapper[4857]: I1201 21:38:30.805217 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2rmn" event={"ID":"41882b8a-22d5-4283-a5e1-8a400181d9de","Type":"ContainerStarted","Data":"c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072"} Dec 01 21:38:31 crc kubenswrapper[4857]: I1201 21:38:31.047201 4857 scope.go:117] "RemoveContainer" containerID="231308b248a8313c68b9b700ade180f803db673c39ee58616469d76955a0adcc" Dec 01 21:38:31 crc kubenswrapper[4857]: I1201 21:38:31.817799 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7x8v8" event={"ID":"0390146f-bbac-4efb-aaef-c7e9c41f8d51","Type":"ContainerStarted","Data":"09373cf0bc85e05e20802c34352030b20b8807f2bbfd4f78f3e752b6c0212cd2"} Dec 01 21:38:31 crc kubenswrapper[4857]: I1201 21:38:31.821058 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thg27" event={"ID":"2235e98c-f4f9-40ae-8f47-788c336e5893","Type":"ContainerStarted","Data":"76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56"} Dec 01 21:38:31 crc kubenswrapper[4857]: I1201 21:38:31.850396 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4e9e374-0931-4274-8bc4-8a9c5e405578" path="/var/lib/kubelet/pods/a4e9e374-0931-4274-8bc4-8a9c5e405578/volumes" Dec 01 21:38:31 crc kubenswrapper[4857]: I1201 21:38:31.858331 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7x8v8" podStartSLOduration=4.668686893 podStartE2EDuration="1m12.858308644s" podCreationTimestamp="2025-12-01 21:37:19 +0000 UTC" firstStartedPulling="2025-12-01 21:37:22.856842015 +0000 UTC m=+161.346904332" lastFinishedPulling="2025-12-01 21:38:31.046463766 +0000 UTC m=+229.536526083" observedRunningTime="2025-12-01 21:38:31.851736586 +0000 UTC m=+230.341798943" watchObservedRunningTime="2025-12-01 21:38:31.858308644 +0000 UTC m=+230.348370961" Dec 01 21:38:31 crc kubenswrapper[4857]: I1201 21:38:31.877816 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r2rmn" podStartSLOduration=5.877435341 podStartE2EDuration="1m12.877795119s" podCreationTimestamp="2025-12-01 21:37:19 +0000 UTC" firstStartedPulling="2025-12-01 21:37:22.761518705 +0000 UTC m=+161.251581012" lastFinishedPulling="2025-12-01 21:38:29.761878463 +0000 UTC m=+228.251940790" observedRunningTime="2025-12-01 21:38:31.877287534 +0000 UTC m=+230.367349851" watchObservedRunningTime="2025-12-01 21:38:31.877795119 +0000 UTC m=+230.367857436" Dec 01 21:38:31 crc kubenswrapper[4857]: I1201 21:38:31.936761 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:38:31 crc kubenswrapper[4857]: I1201 21:38:31.936815 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:38:31 crc kubenswrapper[4857]: I1201 21:38:31.966333 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-thg27" podStartSLOduration=3.771078896 podStartE2EDuration="1m10.966318187s" podCreationTimestamp="2025-12-01 21:37:21 +0000 UTC" firstStartedPulling="2025-12-01 21:37:23.878092872 +0000 UTC m=+162.368155189" lastFinishedPulling="2025-12-01 21:38:31.073332163 +0000 UTC m=+229.563394480" observedRunningTime="2025-12-01 21:38:31.965716699 +0000 UTC m=+230.455779016" watchObservedRunningTime="2025-12-01 21:38:31.966318187 +0000 UTC m=+230.456380504" Dec 01 21:38:31 crc kubenswrapper[4857]: I1201 21:38:31.991492 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4s5mw" podStartSLOduration=7.067105721 podStartE2EDuration="1m13.991476412s" podCreationTimestamp="2025-12-01 21:37:18 +0000 UTC" firstStartedPulling="2025-12-01 21:37:22.793596393 +0000 UTC m=+161.283658710" lastFinishedPulling="2025-12-01 21:38:29.717967074 +0000 UTC m=+228.208029401" observedRunningTime="2025-12-01 21:38:31.989675328 +0000 UTC m=+230.479737645" watchObservedRunningTime="2025-12-01 21:38:31.991476412 +0000 UTC m=+230.481538729" Dec 01 21:38:31 crc kubenswrapper[4857]: I1201 21:38:31.992582 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:38:32 crc kubenswrapper[4857]: I1201 21:38:32.015189 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k9szd" podStartSLOduration=6.027532614 podStartE2EDuration="1m13.015175394s" podCreationTimestamp="2025-12-01 21:37:19 +0000 UTC" firstStartedPulling="2025-12-01 21:37:22.713984454 +0000 UTC m=+161.204046771" lastFinishedPulling="2025-12-01 21:38:29.701627234 +0000 UTC m=+228.191689551" observedRunningTime="2025-12-01 21:38:32.011202665 +0000 UTC m=+230.501264982" watchObservedRunningTime="2025-12-01 21:38:32.015175394 +0000 UTC m=+230.505237711" Dec 01 21:38:32 crc kubenswrapper[4857]: I1201 21:38:32.532398 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:38:32 crc kubenswrapper[4857]: I1201 21:38:32.572126 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:38:32 crc kubenswrapper[4857]: I1201 21:38:32.871447 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:38:37 crc kubenswrapper[4857]: I1201 21:38:37.056005 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntgbq"] Dec 01 21:38:37 crc kubenswrapper[4857]: I1201 21:38:37.057111 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ntgbq" podUID="902b6911-fd58-475c-adb0-22cafe4a2180" containerName="registry-server" containerID="cri-o://36407e50935ef0762111442c54972adb0e91618d5672628075721c7a008d5c6d" gracePeriod=2 Dec 01 21:38:38 crc kubenswrapper[4857]: I1201 21:38:38.870441 4857 generic.go:334] "Generic (PLEG): container finished" podID="902b6911-fd58-475c-adb0-22cafe4a2180" containerID="36407e50935ef0762111442c54972adb0e91618d5672628075721c7a008d5c6d" exitCode=0 Dec 01 21:38:38 crc kubenswrapper[4857]: I1201 21:38:38.870524 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntgbq" event={"ID":"902b6911-fd58-475c-adb0-22cafe4a2180","Type":"ContainerDied","Data":"36407e50935ef0762111442c54972adb0e91618d5672628075721c7a008d5c6d"} Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.404077 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.404564 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.499519 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.693089 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.693152 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.737539 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.767475 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.873199 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p7nl\" (UniqueName: \"kubernetes.io/projected/902b6911-fd58-475c-adb0-22cafe4a2180-kube-api-access-4p7nl\") pod \"902b6911-fd58-475c-adb0-22cafe4a2180\" (UID: \"902b6911-fd58-475c-adb0-22cafe4a2180\") " Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.873273 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/902b6911-fd58-475c-adb0-22cafe4a2180-catalog-content\") pod \"902b6911-fd58-475c-adb0-22cafe4a2180\" (UID: \"902b6911-fd58-475c-adb0-22cafe4a2180\") " Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.873354 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/902b6911-fd58-475c-adb0-22cafe4a2180-utilities\") pod \"902b6911-fd58-475c-adb0-22cafe4a2180\" (UID: \"902b6911-fd58-475c-adb0-22cafe4a2180\") " Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.874382 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/902b6911-fd58-475c-adb0-22cafe4a2180-utilities" (OuterVolumeSpecName: "utilities") pod "902b6911-fd58-475c-adb0-22cafe4a2180" (UID: "902b6911-fd58-475c-adb0-22cafe4a2180"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.883809 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/902b6911-fd58-475c-adb0-22cafe4a2180-kube-api-access-4p7nl" (OuterVolumeSpecName: "kube-api-access-4p7nl") pod "902b6911-fd58-475c-adb0-22cafe4a2180" (UID: "902b6911-fd58-475c-adb0-22cafe4a2180"). InnerVolumeSpecName "kube-api-access-4p7nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.884732 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ntgbq" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.885072 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntgbq" event={"ID":"902b6911-fd58-475c-adb0-22cafe4a2180","Type":"ContainerDied","Data":"6aefb605f627d75803ce20f3ac344f97fbbb1a7fb6b5a0f0c9d7b40a510c6e8b"} Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.885127 4857 scope.go:117] "RemoveContainer" containerID="36407e50935ef0762111442c54972adb0e91618d5672628075721c7a008d5c6d" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.900679 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/902b6911-fd58-475c-adb0-22cafe4a2180-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "902b6911-fd58-475c-adb0-22cafe4a2180" (UID: "902b6911-fd58-475c-adb0-22cafe4a2180"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.936437 4857 scope.go:117] "RemoveContainer" containerID="86d0c1963a159089023764b3ffe77fb660bd9ec52c95613c717daf40a01f943b" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.936593 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.950735 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.969380 4857 scope.go:117] "RemoveContainer" containerID="51c8e3031385eb25a4a560b1474e0b40771a837a44229d576109baddfc72a411" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.975088 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p7nl\" (UniqueName: \"kubernetes.io/projected/902b6911-fd58-475c-adb0-22cafe4a2180-kube-api-access-4p7nl\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.975108 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/902b6911-fd58-475c-adb0-22cafe4a2180-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:39 crc kubenswrapper[4857]: I1201 21:38:39.975118 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/902b6911-fd58-475c-adb0-22cafe4a2180-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:40 crc kubenswrapper[4857]: I1201 21:38:40.232926 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntgbq"] Dec 01 21:38:40 crc kubenswrapper[4857]: I1201 21:38:40.238759 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntgbq"] Dec 01 21:38:40 crc kubenswrapper[4857]: I1201 21:38:40.535515 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:38:40 crc kubenswrapper[4857]: I1201 21:38:40.535589 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:38:40 crc kubenswrapper[4857]: I1201 21:38:40.581872 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:38:40 crc kubenswrapper[4857]: I1201 21:38:40.675067 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:38:40 crc kubenswrapper[4857]: I1201 21:38:40.675136 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:38:40 crc kubenswrapper[4857]: I1201 21:38:40.741590 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:38:40 crc kubenswrapper[4857]: I1201 21:38:40.961071 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:38:40 crc kubenswrapper[4857]: I1201 21:38:40.973420 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:38:41 crc kubenswrapper[4857]: I1201 21:38:41.059669 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r2rmn"] Dec 01 21:38:41 crc kubenswrapper[4857]: I1201 21:38:41.637217 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:38:41 crc kubenswrapper[4857]: I1201 21:38:41.637653 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:38:41 crc kubenswrapper[4857]: I1201 21:38:41.700837 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:38:41 crc kubenswrapper[4857]: I1201 21:38:41.849421 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="902b6911-fd58-475c-adb0-22cafe4a2180" path="/var/lib/kubelet/pods/902b6911-fd58-475c-adb0-22cafe4a2180/volumes" Dec 01 21:38:41 crc kubenswrapper[4857]: I1201 21:38:41.901148 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r2rmn" podUID="41882b8a-22d5-4283-a5e1-8a400181d9de" containerName="registry-server" containerID="cri-o://c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072" gracePeriod=2 Dec 01 21:38:41 crc kubenswrapper[4857]: I1201 21:38:41.958484 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.358225 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.411068 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41882b8a-22d5-4283-a5e1-8a400181d9de-catalog-content\") pod \"41882b8a-22d5-4283-a5e1-8a400181d9de\" (UID: \"41882b8a-22d5-4283-a5e1-8a400181d9de\") " Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.411163 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41882b8a-22d5-4283-a5e1-8a400181d9de-utilities\") pod \"41882b8a-22d5-4283-a5e1-8a400181d9de\" (UID: \"41882b8a-22d5-4283-a5e1-8a400181d9de\") " Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.411234 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2htwv\" (UniqueName: \"kubernetes.io/projected/41882b8a-22d5-4283-a5e1-8a400181d9de-kube-api-access-2htwv\") pod \"41882b8a-22d5-4283-a5e1-8a400181d9de\" (UID: \"41882b8a-22d5-4283-a5e1-8a400181d9de\") " Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.412515 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41882b8a-22d5-4283-a5e1-8a400181d9de-utilities" (OuterVolumeSpecName: "utilities") pod "41882b8a-22d5-4283-a5e1-8a400181d9de" (UID: "41882b8a-22d5-4283-a5e1-8a400181d9de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.424495 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41882b8a-22d5-4283-a5e1-8a400181d9de-kube-api-access-2htwv" (OuterVolumeSpecName: "kube-api-access-2htwv") pod "41882b8a-22d5-4283-a5e1-8a400181d9de" (UID: "41882b8a-22d5-4283-a5e1-8a400181d9de"). InnerVolumeSpecName "kube-api-access-2htwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.453644 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41882b8a-22d5-4283-a5e1-8a400181d9de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41882b8a-22d5-4283-a5e1-8a400181d9de" (UID: "41882b8a-22d5-4283-a5e1-8a400181d9de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.513122 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41882b8a-22d5-4283-a5e1-8a400181d9de-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.513194 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41882b8a-22d5-4283-a5e1-8a400181d9de-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.513215 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2htwv\" (UniqueName: \"kubernetes.io/projected/41882b8a-22d5-4283-a5e1-8a400181d9de-kube-api-access-2htwv\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.915521 4857 generic.go:334] "Generic (PLEG): container finished" podID="41882b8a-22d5-4283-a5e1-8a400181d9de" containerID="c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072" exitCode=0 Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.916994 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2rmn" event={"ID":"41882b8a-22d5-4283-a5e1-8a400181d9de","Type":"ContainerDied","Data":"c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072"} Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.917063 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2rmn" event={"ID":"41882b8a-22d5-4283-a5e1-8a400181d9de","Type":"ContainerDied","Data":"e28a86b5fef854c914b3a8293c617c66ab1d6680a1e871ebaaf5836600386eef"} Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.917072 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2rmn" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.917089 4857 scope.go:117] "RemoveContainer" containerID="c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.935633 4857 scope.go:117] "RemoveContainer" containerID="2762ba9da4f468a30fffc2c8a642df5491bba0db3c30e40aa733ddeae717c6dd" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.956873 4857 scope.go:117] "RemoveContainer" containerID="5cfc2f8fa4a7d4883db82d8dd92c89440dee18dd62d7bdc44314387953cb0717" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.959653 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r2rmn"] Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.965904 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r2rmn"] Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.979822 4857 scope.go:117] "RemoveContainer" containerID="c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072" Dec 01 21:38:42 crc kubenswrapper[4857]: E1201 21:38:42.980313 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072\": container with ID starting with c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072 not found: ID does not exist" containerID="c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.980359 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072"} err="failed to get container status \"c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072\": rpc error: code = NotFound desc = could not find container \"c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072\": container with ID starting with c4897f029452f0abbe127a0d221c4b8948bcd66167359a1b68e5b4f04d657072 not found: ID does not exist" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.980389 4857 scope.go:117] "RemoveContainer" containerID="2762ba9da4f468a30fffc2c8a642df5491bba0db3c30e40aa733ddeae717c6dd" Dec 01 21:38:42 crc kubenswrapper[4857]: E1201 21:38:42.980766 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2762ba9da4f468a30fffc2c8a642df5491bba0db3c30e40aa733ddeae717c6dd\": container with ID starting with 2762ba9da4f468a30fffc2c8a642df5491bba0db3c30e40aa733ddeae717c6dd not found: ID does not exist" containerID="2762ba9da4f468a30fffc2c8a642df5491bba0db3c30e40aa733ddeae717c6dd" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.980790 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2762ba9da4f468a30fffc2c8a642df5491bba0db3c30e40aa733ddeae717c6dd"} err="failed to get container status \"2762ba9da4f468a30fffc2c8a642df5491bba0db3c30e40aa733ddeae717c6dd\": rpc error: code = NotFound desc = could not find container \"2762ba9da4f468a30fffc2c8a642df5491bba0db3c30e40aa733ddeae717c6dd\": container with ID starting with 2762ba9da4f468a30fffc2c8a642df5491bba0db3c30e40aa733ddeae717c6dd not found: ID does not exist" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.980806 4857 scope.go:117] "RemoveContainer" containerID="5cfc2f8fa4a7d4883db82d8dd92c89440dee18dd62d7bdc44314387953cb0717" Dec 01 21:38:42 crc kubenswrapper[4857]: E1201 21:38:42.981211 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cfc2f8fa4a7d4883db82d8dd92c89440dee18dd62d7bdc44314387953cb0717\": container with ID starting with 5cfc2f8fa4a7d4883db82d8dd92c89440dee18dd62d7bdc44314387953cb0717 not found: ID does not exist" containerID="5cfc2f8fa4a7d4883db82d8dd92c89440dee18dd62d7bdc44314387953cb0717" Dec 01 21:38:42 crc kubenswrapper[4857]: I1201 21:38:42.981250 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cfc2f8fa4a7d4883db82d8dd92c89440dee18dd62d7bdc44314387953cb0717"} err="failed to get container status \"5cfc2f8fa4a7d4883db82d8dd92c89440dee18dd62d7bdc44314387953cb0717\": rpc error: code = NotFound desc = could not find container \"5cfc2f8fa4a7d4883db82d8dd92c89440dee18dd62d7bdc44314387953cb0717\": container with ID starting with 5cfc2f8fa4a7d4883db82d8dd92c89440dee18dd62d7bdc44314387953cb0717 not found: ID does not exist" Dec 01 21:38:43 crc kubenswrapper[4857]: I1201 21:38:43.846271 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41882b8a-22d5-4283-a5e1-8a400181d9de" path="/var/lib/kubelet/pods/41882b8a-22d5-4283-a5e1-8a400181d9de/volumes" Dec 01 21:38:44 crc kubenswrapper[4857]: I1201 21:38:44.066635 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7x8v8"] Dec 01 21:38:44 crc kubenswrapper[4857]: I1201 21:38:44.067139 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7x8v8" podUID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" containerName="registry-server" containerID="cri-o://09373cf0bc85e05e20802c34352030b20b8807f2bbfd4f78f3e752b6c0212cd2" gracePeriod=2 Dec 01 21:38:44 crc kubenswrapper[4857]: E1201 21:38:44.209959 4857 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0390146f_bbac_4efb_aaef_c7e9c41f8d51.slice/crio-09373cf0bc85e05e20802c34352030b20b8807f2bbfd4f78f3e752b6c0212cd2.scope\": RecentStats: unable to find data in memory cache]" Dec 01 21:38:44 crc kubenswrapper[4857]: I1201 21:38:44.941345 4857 generic.go:334] "Generic (PLEG): container finished" podID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" containerID="09373cf0bc85e05e20802c34352030b20b8807f2bbfd4f78f3e752b6c0212cd2" exitCode=0 Dec 01 21:38:44 crc kubenswrapper[4857]: I1201 21:38:44.941394 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7x8v8" event={"ID":"0390146f-bbac-4efb-aaef-c7e9c41f8d51","Type":"ContainerDied","Data":"09373cf0bc85e05e20802c34352030b20b8807f2bbfd4f78f3e752b6c0212cd2"} Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.084530 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.151475 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6zrm\" (UniqueName: \"kubernetes.io/projected/0390146f-bbac-4efb-aaef-c7e9c41f8d51-kube-api-access-q6zrm\") pod \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\" (UID: \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\") " Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.151513 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0390146f-bbac-4efb-aaef-c7e9c41f8d51-catalog-content\") pod \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\" (UID: \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\") " Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.151545 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0390146f-bbac-4efb-aaef-c7e9c41f8d51-utilities\") pod \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\" (UID: \"0390146f-bbac-4efb-aaef-c7e9c41f8d51\") " Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.152439 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0390146f-bbac-4efb-aaef-c7e9c41f8d51-utilities" (OuterVolumeSpecName: "utilities") pod "0390146f-bbac-4efb-aaef-c7e9c41f8d51" (UID: "0390146f-bbac-4efb-aaef-c7e9c41f8d51"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.156327 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0390146f-bbac-4efb-aaef-c7e9c41f8d51-kube-api-access-q6zrm" (OuterVolumeSpecName: "kube-api-access-q6zrm") pod "0390146f-bbac-4efb-aaef-c7e9c41f8d51" (UID: "0390146f-bbac-4efb-aaef-c7e9c41f8d51"). InnerVolumeSpecName "kube-api-access-q6zrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.219337 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0390146f-bbac-4efb-aaef-c7e9c41f8d51-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0390146f-bbac-4efb-aaef-c7e9c41f8d51" (UID: "0390146f-bbac-4efb-aaef-c7e9c41f8d51"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.252870 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6zrm\" (UniqueName: \"kubernetes.io/projected/0390146f-bbac-4efb-aaef-c7e9c41f8d51-kube-api-access-q6zrm\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.252895 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0390146f-bbac-4efb-aaef-c7e9c41f8d51-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.252904 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0390146f-bbac-4efb-aaef-c7e9c41f8d51-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.566432 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" podUID="7ff6afde-262c-4ebf-be69-2e043814271c" containerName="oauth-openshift" containerID="cri-o://0cdc776a88afcca486adf56ecc6ec9b4d2b0cd60649a605fdf9766f0524322cd" gracePeriod=15 Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.953421 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7x8v8" event={"ID":"0390146f-bbac-4efb-aaef-c7e9c41f8d51","Type":"ContainerDied","Data":"70e10783aba9cd279cb92faab8b281e264ad9939ad04b08bb7a8976a1b6a1583"} Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.953479 4857 scope.go:117] "RemoveContainer" containerID="09373cf0bc85e05e20802c34352030b20b8807f2bbfd4f78f3e752b6c0212cd2" Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.953578 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7x8v8" Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.955682 4857 generic.go:334] "Generic (PLEG): container finished" podID="7ff6afde-262c-4ebf-be69-2e043814271c" containerID="0cdc776a88afcca486adf56ecc6ec9b4d2b0cd60649a605fdf9766f0524322cd" exitCode=0 Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.955729 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" event={"ID":"7ff6afde-262c-4ebf-be69-2e043814271c","Type":"ContainerDied","Data":"0cdc776a88afcca486adf56ecc6ec9b4d2b0cd60649a605fdf9766f0524322cd"} Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.983505 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7x8v8"] Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.984643 4857 scope.go:117] "RemoveContainer" containerID="daa9a03301e1e58c40db0e87f032767563ebe48b57e46da1dd0754fc91eb7329" Dec 01 21:38:45 crc kubenswrapper[4857]: I1201 21:38:45.986513 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7x8v8"] Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.015492 4857 scope.go:117] "RemoveContainer" containerID="2d83cbf2bc6f2d24c3943aaf996c9c9d0cd3810cace60b3f1f3ace8987ace44d" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.077074 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.167315 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ff6afde-262c-4ebf-be69-2e043814271c-audit-dir\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.167425 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-trusted-ca-bundle\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.167637 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ff6afde-262c-4ebf-be69-2e043814271c-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.168699 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-serving-cert\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.169063 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.169109 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-session\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.169220 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-audit-policies\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.169428 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-router-certs\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.169524 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-ocp-branding-template\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.169608 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-error\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.169685 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-idp-0-file-data\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.169728 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-service-ca\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.169902 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsknc\" (UniqueName: \"kubernetes.io/projected/7ff6afde-262c-4ebf-be69-2e043814271c-kube-api-access-hsknc\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.169997 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-cliconfig\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.170035 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-login\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.170107 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-provider-selection\") pod \"7ff6afde-262c-4ebf-be69-2e043814271c\" (UID: \"7ff6afde-262c-4ebf-be69-2e043814271c\") " Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.170596 4857 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ff6afde-262c-4ebf-be69-2e043814271c-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.170501 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.170633 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.171117 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.172398 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.174345 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.174858 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.175894 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.177175 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.178809 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.179577 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ff6afde-262c-4ebf-be69-2e043814271c-kube-api-access-hsknc" (OuterVolumeSpecName: "kube-api-access-hsknc") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "kube-api-access-hsknc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.180752 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.190241 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.190683 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7ff6afde-262c-4ebf-be69-2e043814271c" (UID: "7ff6afde-262c-4ebf-be69-2e043814271c"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.272762 4857 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.272814 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.272838 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.272862 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.272882 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.272902 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.272922 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsknc\" (UniqueName: \"kubernetes.io/projected/7ff6afde-262c-4ebf-be69-2e043814271c-kube-api-access-hsknc\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.272941 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.272959 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.272981 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.273001 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.273025 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ff6afde-262c-4ebf-be69-2e043814271c-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.962137 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.962145 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mmb9s" event={"ID":"7ff6afde-262c-4ebf-be69-2e043814271c","Type":"ContainerDied","Data":"3e2a1dc94974b96496e9e77a178d31736c515dc08b50c015a61647b4267cda17"} Dec 01 21:38:46 crc kubenswrapper[4857]: I1201 21:38:46.962245 4857 scope.go:117] "RemoveContainer" containerID="0cdc776a88afcca486adf56ecc6ec9b4d2b0cd60649a605fdf9766f0524322cd" Dec 01 21:38:47 crc kubenswrapper[4857]: I1201 21:38:47.001962 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mmb9s"] Dec 01 21:38:47 crc kubenswrapper[4857]: I1201 21:38:47.005681 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mmb9s"] Dec 01 21:38:47 crc kubenswrapper[4857]: I1201 21:38:47.845913 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" path="/var/lib/kubelet/pods/0390146f-bbac-4efb-aaef-c7e9c41f8d51/volumes" Dec 01 21:38:47 crc kubenswrapper[4857]: I1201 21:38:47.847337 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ff6afde-262c-4ebf-be69-2e043814271c" path="/var/lib/kubelet/pods/7ff6afde-262c-4ebf-be69-2e043814271c/volumes" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.375031 4857 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376172 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff6afde-262c-4ebf-be69-2e043814271c" containerName="oauth-openshift" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376199 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff6afde-262c-4ebf-be69-2e043814271c" containerName="oauth-openshift" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376224 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="902b6911-fd58-475c-adb0-22cafe4a2180" containerName="extract-utilities" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376244 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="902b6911-fd58-475c-adb0-22cafe4a2180" containerName="extract-utilities" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376272 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" containerName="registry-server" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376289 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" containerName="registry-server" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376313 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" containerName="extract-utilities" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376327 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" containerName="extract-utilities" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376350 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="902b6911-fd58-475c-adb0-22cafe4a2180" containerName="extract-content" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376365 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="902b6911-fd58-475c-adb0-22cafe4a2180" containerName="extract-content" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376389 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41882b8a-22d5-4283-a5e1-8a400181d9de" containerName="registry-server" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376407 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="41882b8a-22d5-4283-a5e1-8a400181d9de" containerName="registry-server" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376425 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e9e374-0931-4274-8bc4-8a9c5e405578" containerName="registry-server" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376437 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e9e374-0931-4274-8bc4-8a9c5e405578" containerName="registry-server" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376455 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" containerName="extract-content" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376469 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" containerName="extract-content" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376486 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e9e374-0931-4274-8bc4-8a9c5e405578" containerName="extract-utilities" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376498 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e9e374-0931-4274-8bc4-8a9c5e405578" containerName="extract-utilities" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376520 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e9e374-0931-4274-8bc4-8a9c5e405578" containerName="extract-content" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376535 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e9e374-0931-4274-8bc4-8a9c5e405578" containerName="extract-content" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376555 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41882b8a-22d5-4283-a5e1-8a400181d9de" containerName="extract-utilities" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376571 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="41882b8a-22d5-4283-a5e1-8a400181d9de" containerName="extract-utilities" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376646 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9765efe-fedf-4dc9-b28f-095fedff216e" containerName="pruner" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376665 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9765efe-fedf-4dc9-b28f-095fedff216e" containerName="pruner" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376687 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41882b8a-22d5-4283-a5e1-8a400181d9de" containerName="extract-content" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376703 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="41882b8a-22d5-4283-a5e1-8a400181d9de" containerName="extract-content" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.376729 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="902b6911-fd58-475c-adb0-22cafe4a2180" containerName="registry-server" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.376746 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="902b6911-fd58-475c-adb0-22cafe4a2180" containerName="registry-server" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.377085 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ff6afde-262c-4ebf-be69-2e043814271c" containerName="oauth-openshift" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.377123 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4e9e374-0931-4274-8bc4-8a9c5e405578" containerName="registry-server" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.377145 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="902b6911-fd58-475c-adb0-22cafe4a2180" containerName="registry-server" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.377173 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9765efe-fedf-4dc9-b28f-095fedff216e" containerName="pruner" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.377196 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="0390146f-bbac-4efb-aaef-c7e9c41f8d51" containerName="registry-server" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.377221 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="41882b8a-22d5-4283-a5e1-8a400181d9de" containerName="registry-server" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.377955 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.422360 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.443793 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.443873 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.443917 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.444269 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.444331 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.456703 4857 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.457151 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16" gracePeriod=15 Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.457369 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773" gracePeriod=15 Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.457461 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14" gracePeriod=15 Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.457627 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091" gracePeriod=15 Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.457741 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda" gracePeriod=15 Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.459097 4857 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.459829 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.459852 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.459870 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.459922 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.459943 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.459955 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.460005 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.460020 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.460082 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.460096 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.460110 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.460122 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.460420 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.460479 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.460500 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.460526 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.460541 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546029 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546107 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546147 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546184 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546263 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546311 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546343 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546378 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546484 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546536 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546571 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546604 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.546652 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.610571 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:38:50Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:38:50Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:38:50Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:38:50Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:0a91c078668b35ed5bcf52d416f9e34cd978491b3668ee1c86b7b5d07164060e\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:f77db38a1bcd07abf8cd56276024cc7ab24cf70a5b8f7e39fb08c8862dee0785\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1607737696},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:61d26ea7a26593e06b479225c8fc57ff8d82b6d27076fb80529751fe70b54a64\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:bb24465fd94d9154042e9a420d8ddc1a6ec16f1e59158f9b8c99df7f98dd3029\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1203933014},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:5c6aa673c1d438af878bee73db58f30a1a660ab640741133ecad7b9114a866d7\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ae051258f5d538fc822b2a546b73657b380190388fdbf6a1ebe21a7e08c43ad\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201251152},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:e8990432556acad31519b1a73ec32f32d27c2034cf9e5cc4db8980efc7331594\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:ebe9f523f5c211a3a0f2570331dddcd5be15b12c1fecd9b8b121f881bfaad029\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1129027903},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.611209 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.611568 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.611892 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.612126 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.612144 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.647832 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.647976 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.648213 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.648365 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.648263 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.648410 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.713841 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:38:50 crc kubenswrapper[4857]: E1201 21:38:50.735376 4857 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.136:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d3534ff2e2701 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 21:38:50.734888705 +0000 UTC m=+249.224951022,LastTimestamp:2025-12-01 21:38:50.734888705 +0000 UTC m=+249.224951022,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.754481 4857 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.754658 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.993011 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a7bbf941505864617f3256cfcbabeead7ab714914232940fd0c58d1ac74d019b"} Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.996378 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.997214 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773" exitCode=0 Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.997349 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14" exitCode=0 Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.997445 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091" exitCode=0 Dec 01 21:38:50 crc kubenswrapper[4857]: I1201 21:38:50.997532 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda" exitCode=2 Dec 01 21:38:51 crc kubenswrapper[4857]: I1201 21:38:51.839404 4857 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:51 crc kubenswrapper[4857]: I1201 21:38:51.840724 4857 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.007274 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"db076f2683fbbb30d0b00304714478cf038f8a8272c27b3bec2b83b75bd8c928"} Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.008496 4857 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:52 crc kubenswrapper[4857]: E1201 21:38:52.861573 4857 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.136:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" volumeName="registry-storage" Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.939335 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.940983 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.941971 4857 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.942663 4857 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.982025 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.982299 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.982405 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.982472 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.982566 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.982862 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.982896 4857 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:52 crc kubenswrapper[4857]: I1201 21:38:52.983362 4857 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.021286 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.022723 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16" exitCode=0 Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.022805 4857 scope.go:117] "RemoveContainer" containerID="0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.022847 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.041551 4857 scope.go:117] "RemoveContainer" containerID="1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.053687 4857 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.054538 4857 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.064810 4857 scope.go:117] "RemoveContainer" containerID="d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.084780 4857 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.085638 4857 scope.go:117] "RemoveContainer" containerID="36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.112950 4857 scope.go:117] "RemoveContainer" containerID="335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.138197 4857 scope.go:117] "RemoveContainer" containerID="e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.167714 4857 scope.go:117] "RemoveContainer" containerID="0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773" Dec 01 21:38:53 crc kubenswrapper[4857]: E1201 21:38:53.168458 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\": container with ID starting with 0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773 not found: ID does not exist" containerID="0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.168537 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773"} err="failed to get container status \"0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\": rpc error: code = NotFound desc = could not find container \"0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773\": container with ID starting with 0a8ba7cb90975e23e4a72c4efb6447d895aa31929857709c937f730bce848773 not found: ID does not exist" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.168595 4857 scope.go:117] "RemoveContainer" containerID="1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14" Dec 01 21:38:53 crc kubenswrapper[4857]: E1201 21:38:53.169360 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\": container with ID starting with 1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14 not found: ID does not exist" containerID="1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.169418 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14"} err="failed to get container status \"1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\": rpc error: code = NotFound desc = could not find container \"1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14\": container with ID starting with 1272fa8d52c1d4f133df67b7a96b2fdce886182367fdad43fc67aff6d5741b14 not found: ID does not exist" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.169460 4857 scope.go:117] "RemoveContainer" containerID="d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091" Dec 01 21:38:53 crc kubenswrapper[4857]: E1201 21:38:53.170236 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\": container with ID starting with d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091 not found: ID does not exist" containerID="d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.170326 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091"} err="failed to get container status \"d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\": rpc error: code = NotFound desc = could not find container \"d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091\": container with ID starting with d9d9465944498581a060d86ea8a18b7f955b358dc701e658969db836264b3091 not found: ID does not exist" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.170386 4857 scope.go:117] "RemoveContainer" containerID="36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda" Dec 01 21:38:53 crc kubenswrapper[4857]: E1201 21:38:53.171235 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\": container with ID starting with 36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda not found: ID does not exist" containerID="36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.171298 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda"} err="failed to get container status \"36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\": rpc error: code = NotFound desc = could not find container \"36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda\": container with ID starting with 36b542cdf140f38974c45a84ae0e64c16c4aa862a1c0603b6859ab4349f01fda not found: ID does not exist" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.171331 4857 scope.go:117] "RemoveContainer" containerID="335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16" Dec 01 21:38:53 crc kubenswrapper[4857]: E1201 21:38:53.171766 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\": container with ID starting with 335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16 not found: ID does not exist" containerID="335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.171808 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16"} err="failed to get container status \"335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\": rpc error: code = NotFound desc = could not find container \"335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16\": container with ID starting with 335846a2c6beea7b55661214d8ca9ea5781ee0bb65f866f099352dddcff7ec16 not found: ID does not exist" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.171833 4857 scope.go:117] "RemoveContainer" containerID="e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230" Dec 01 21:38:53 crc kubenswrapper[4857]: E1201 21:38:53.172355 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\": container with ID starting with e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230 not found: ID does not exist" containerID="e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.172396 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230"} err="failed to get container status \"e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\": rpc error: code = NotFound desc = could not find container \"e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230\": container with ID starting with e16b6f5577e1c458cab7df0b54320c284a8a654b54a25e4b267b1d35cace8230 not found: ID does not exist" Dec 01 21:38:53 crc kubenswrapper[4857]: I1201 21:38:53.854949 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 01 21:38:54 crc kubenswrapper[4857]: E1201 21:38:54.853443 4857 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:54 crc kubenswrapper[4857]: E1201 21:38:54.854185 4857 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:54 crc kubenswrapper[4857]: E1201 21:38:54.854677 4857 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:54 crc kubenswrapper[4857]: E1201 21:38:54.855073 4857 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:54 crc kubenswrapper[4857]: E1201 21:38:54.855506 4857 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:54 crc kubenswrapper[4857]: I1201 21:38:54.855564 4857 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 01 21:38:54 crc kubenswrapper[4857]: E1201 21:38:54.856017 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="200ms" Dec 01 21:38:55 crc kubenswrapper[4857]: E1201 21:38:55.057821 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="400ms" Dec 01 21:38:55 crc kubenswrapper[4857]: E1201 21:38:55.460536 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="800ms" Dec 01 21:38:56 crc kubenswrapper[4857]: I1201 21:38:56.046724 4857 generic.go:334] "Generic (PLEG): container finished" podID="c350e136-f264-40ea-9b0e-9211dea68129" containerID="d943040e60619af528fafac613ed9807b10f7f1145b385ab05950957ddb1d225" exitCode=0 Dec 01 21:38:56 crc kubenswrapper[4857]: I1201 21:38:56.046818 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c350e136-f264-40ea-9b0e-9211dea68129","Type":"ContainerDied","Data":"d943040e60619af528fafac613ed9807b10f7f1145b385ab05950957ddb1d225"} Dec 01 21:38:56 crc kubenswrapper[4857]: I1201 21:38:56.048223 4857 status_manager.go:851] "Failed to get status for pod" podUID="c350e136-f264-40ea-9b0e-9211dea68129" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:56 crc kubenswrapper[4857]: I1201 21:38:56.048821 4857 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:56 crc kubenswrapper[4857]: E1201 21:38:56.263337 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="1.6s" Dec 01 21:38:57 crc kubenswrapper[4857]: I1201 21:38:57.411180 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:57 crc kubenswrapper[4857]: I1201 21:38:57.412183 4857 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:57 crc kubenswrapper[4857]: I1201 21:38:57.412697 4857 status_manager.go:851] "Failed to get status for pod" podUID="c350e136-f264-40ea-9b0e-9211dea68129" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:57 crc kubenswrapper[4857]: I1201 21:38:57.447679 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c350e136-f264-40ea-9b0e-9211dea68129-kubelet-dir\") pod \"c350e136-f264-40ea-9b0e-9211dea68129\" (UID: \"c350e136-f264-40ea-9b0e-9211dea68129\") " Dec 01 21:38:57 crc kubenswrapper[4857]: I1201 21:38:57.447798 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c350e136-f264-40ea-9b0e-9211dea68129-kube-api-access\") pod \"c350e136-f264-40ea-9b0e-9211dea68129\" (UID: \"c350e136-f264-40ea-9b0e-9211dea68129\") " Dec 01 21:38:57 crc kubenswrapper[4857]: I1201 21:38:57.447825 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c350e136-f264-40ea-9b0e-9211dea68129-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c350e136-f264-40ea-9b0e-9211dea68129" (UID: "c350e136-f264-40ea-9b0e-9211dea68129"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:38:57 crc kubenswrapper[4857]: I1201 21:38:57.447958 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c350e136-f264-40ea-9b0e-9211dea68129-var-lock" (OuterVolumeSpecName: "var-lock") pod "c350e136-f264-40ea-9b0e-9211dea68129" (UID: "c350e136-f264-40ea-9b0e-9211dea68129"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:38:57 crc kubenswrapper[4857]: I1201 21:38:57.448410 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c350e136-f264-40ea-9b0e-9211dea68129-var-lock\") pod \"c350e136-f264-40ea-9b0e-9211dea68129\" (UID: \"c350e136-f264-40ea-9b0e-9211dea68129\") " Dec 01 21:38:57 crc kubenswrapper[4857]: I1201 21:38:57.448909 4857 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c350e136-f264-40ea-9b0e-9211dea68129-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:57 crc kubenswrapper[4857]: I1201 21:38:57.448930 4857 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c350e136-f264-40ea-9b0e-9211dea68129-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:57 crc kubenswrapper[4857]: I1201 21:38:57.455842 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c350e136-f264-40ea-9b0e-9211dea68129-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c350e136-f264-40ea-9b0e-9211dea68129" (UID: "c350e136-f264-40ea-9b0e-9211dea68129"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:38:57 crc kubenswrapper[4857]: I1201 21:38:57.550082 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c350e136-f264-40ea-9b0e-9211dea68129-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 21:38:57 crc kubenswrapper[4857]: E1201 21:38:57.865802 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="3.2s" Dec 01 21:38:58 crc kubenswrapper[4857]: I1201 21:38:58.063305 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c350e136-f264-40ea-9b0e-9211dea68129","Type":"ContainerDied","Data":"3db2375a7f1ae95b8d53a8cb13468d5fabb6243dc2e7745834ae187db9a5d774"} Dec 01 21:38:58 crc kubenswrapper[4857]: I1201 21:38:58.063369 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3db2375a7f1ae95b8d53a8cb13468d5fabb6243dc2e7745834ae187db9a5d774" Dec 01 21:38:58 crc kubenswrapper[4857]: I1201 21:38:58.063378 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 21:38:58 crc kubenswrapper[4857]: I1201 21:38:58.069464 4857 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:38:58 crc kubenswrapper[4857]: I1201 21:38:58.070572 4857 status_manager.go:851] "Failed to get status for pod" podUID="c350e136-f264-40ea-9b0e-9211dea68129" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:00 crc kubenswrapper[4857]: E1201 21:39:00.584071 4857 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.136:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d3534ff2e2701 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 21:38:50.734888705 +0000 UTC m=+249.224951022,LastTimestamp:2025-12-01 21:38:50.734888705 +0000 UTC m=+249.224951022,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 21:39:00 crc kubenswrapper[4857]: E1201 21:39:00.774543 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:39:00Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:39:00Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:39:00Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T21:39:00Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:0a91c078668b35ed5bcf52d416f9e34cd978491b3668ee1c86b7b5d07164060e\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:f77db38a1bcd07abf8cd56276024cc7ab24cf70a5b8f7e39fb08c8862dee0785\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1607737696},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:61d26ea7a26593e06b479225c8fc57ff8d82b6d27076fb80529751fe70b54a64\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:bb24465fd94d9154042e9a420d8ddc1a6ec16f1e59158f9b8c99df7f98dd3029\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1203933014},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:5c6aa673c1d438af878bee73db58f30a1a660ab640741133ecad7b9114a866d7\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ae051258f5d538fc822b2a546b73657b380190388fdbf6a1ebe21a7e08c43ad\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201251152},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:e8990432556acad31519b1a73ec32f32d27c2034cf9e5cc4db8980efc7331594\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:ebe9f523f5c211a3a0f2570331dddcd5be15b12c1fecd9b8b121f881bfaad029\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1129027903},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:00 crc kubenswrapper[4857]: E1201 21:39:00.775123 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:00 crc kubenswrapper[4857]: E1201 21:39:00.775554 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:00 crc kubenswrapper[4857]: E1201 21:39:00.775891 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:00 crc kubenswrapper[4857]: E1201 21:39:00.776251 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:00 crc kubenswrapper[4857]: E1201 21:39:00.776286 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 21:39:01 crc kubenswrapper[4857]: E1201 21:39:01.067281 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="6.4s" Dec 01 21:39:01 crc kubenswrapper[4857]: I1201 21:39:01.841498 4857 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:01 crc kubenswrapper[4857]: I1201 21:39:01.842121 4857 status_manager.go:851] "Failed to get status for pod" podUID="c350e136-f264-40ea-9b0e-9211dea68129" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:03 crc kubenswrapper[4857]: I1201 21:39:03.834392 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:39:03 crc kubenswrapper[4857]: I1201 21:39:03.835725 4857 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:03 crc kubenswrapper[4857]: I1201 21:39:03.836339 4857 status_manager.go:851] "Failed to get status for pod" podUID="c350e136-f264-40ea-9b0e-9211dea68129" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:03 crc kubenswrapper[4857]: I1201 21:39:03.857868 4857 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:03 crc kubenswrapper[4857]: I1201 21:39:03.857908 4857 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:03 crc kubenswrapper[4857]: E1201 21:39:03.858551 4857 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:39:03 crc kubenswrapper[4857]: I1201 21:39:03.859362 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:39:04 crc kubenswrapper[4857]: I1201 21:39:04.104470 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"48ce5573dd53c8bb2f937a24cd81e08d41279d2338b971b5820fae9e44137d7d"} Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.115750 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.116307 4857 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc" exitCode=1 Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.116436 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc"} Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.117325 4857 scope.go:117] "RemoveContainer" containerID="f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc" Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.118297 4857 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.119022 4857 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.119520 4857 status_manager.go:851] "Failed to get status for pod" podUID="c350e136-f264-40ea-9b0e-9211dea68129" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.121488 4857 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="86093639c61540aa3e431003478f2a97a40b5defa0eb68c274999bdc728904a9" exitCode=0 Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.121548 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"86093639c61540aa3e431003478f2a97a40b5defa0eb68c274999bdc728904a9"} Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.121916 4857 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.121949 4857 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:05 crc kubenswrapper[4857]: E1201 21:39:05.122402 4857 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.122753 4857 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.123271 4857 status_manager.go:851] "Failed to get status for pod" podUID="c350e136-f264-40ea-9b0e-9211dea68129" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.123922 4857 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Dec 01 21:39:05 crc kubenswrapper[4857]: I1201 21:39:05.775841 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:39:06 crc kubenswrapper[4857]: I1201 21:39:06.131800 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 21:39:06 crc kubenswrapper[4857]: I1201 21:39:06.131929 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d600387d4b3fb1cfe43bd2b6cc1396aa79d3be00e6ce66722e0cbd9ea31e9a2b"} Dec 01 21:39:06 crc kubenswrapper[4857]: I1201 21:39:06.134947 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"250572fda1ba954d8158f7f3c5d553a0d5c6df2cdc39354e477b158c64f03535"} Dec 01 21:39:06 crc kubenswrapper[4857]: I1201 21:39:06.135015 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6bce748ea95c90f8a7be448310725c0640ec8b5f54783ca11b1a695e25c37e05"} Dec 01 21:39:06 crc kubenswrapper[4857]: I1201 21:39:06.135036 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"040b11eb53397fdf9f7f604d50dae10b1dbe9e2f5c8b2dea74f81042d88039a1"} Dec 01 21:39:07 crc kubenswrapper[4857]: I1201 21:39:07.141349 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"39e9dd07be2cc480193d56a40ebaf0ae9df1768da30317fc64a7fd7d64e62c0d"} Dec 01 21:39:07 crc kubenswrapper[4857]: I1201 21:39:07.141621 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"853f841902f8ba61f8a2213588b78014a3a935669bfc3902ad5963ed778c4665"} Dec 01 21:39:07 crc kubenswrapper[4857]: I1201 21:39:07.141714 4857 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:07 crc kubenswrapper[4857]: I1201 21:39:07.141739 4857 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:07 crc kubenswrapper[4857]: I1201 21:39:07.863520 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:39:08 crc kubenswrapper[4857]: I1201 21:39:08.859924 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:39:08 crc kubenswrapper[4857]: I1201 21:39:08.860306 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:39:08 crc kubenswrapper[4857]: I1201 21:39:08.868010 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:39:12 crc kubenswrapper[4857]: I1201 21:39:12.164738 4857 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:39:13 crc kubenswrapper[4857]: I1201 21:39:13.175526 4857 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:13 crc kubenswrapper[4857]: I1201 21:39:13.175872 4857 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:13 crc kubenswrapper[4857]: I1201 21:39:13.175561 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:39:13 crc kubenswrapper[4857]: I1201 21:39:13.184663 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:39:13 crc kubenswrapper[4857]: I1201 21:39:13.188480 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="e6e32e19-989e-46d3-a6af-59cff3147e3d" Dec 01 21:39:14 crc kubenswrapper[4857]: I1201 21:39:14.182915 4857 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:14 crc kubenswrapper[4857]: I1201 21:39:14.182973 4857 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:15 crc kubenswrapper[4857]: I1201 21:39:15.190522 4857 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:15 crc kubenswrapper[4857]: I1201 21:39:15.190942 4857 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:15 crc kubenswrapper[4857]: I1201 21:39:15.775667 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:39:15 crc kubenswrapper[4857]: I1201 21:39:15.775989 4857 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 01 21:39:15 crc kubenswrapper[4857]: I1201 21:39:15.776135 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 01 21:39:21 crc kubenswrapper[4857]: I1201 21:39:21.813266 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 21:39:21 crc kubenswrapper[4857]: I1201 21:39:21.856101 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="e6e32e19-989e-46d3-a6af-59cff3147e3d" Dec 01 21:39:21 crc kubenswrapper[4857]: I1201 21:39:21.949653 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 21:39:22 crc kubenswrapper[4857]: I1201 21:39:22.638063 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 21:39:22 crc kubenswrapper[4857]: I1201 21:39:22.725584 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 21:39:23 crc kubenswrapper[4857]: I1201 21:39:23.344909 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 21:39:23 crc kubenswrapper[4857]: I1201 21:39:23.446551 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 21:39:23 crc kubenswrapper[4857]: I1201 21:39:23.510609 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 21:39:23 crc kubenswrapper[4857]: I1201 21:39:23.572706 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 21:39:23 crc kubenswrapper[4857]: I1201 21:39:23.585111 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 21:39:23 crc kubenswrapper[4857]: I1201 21:39:23.862335 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 21:39:23 crc kubenswrapper[4857]: I1201 21:39:23.897876 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 21:39:23 crc kubenswrapper[4857]: I1201 21:39:23.986288 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 21:39:24 crc kubenswrapper[4857]: I1201 21:39:24.048588 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 21:39:24 crc kubenswrapper[4857]: I1201 21:39:24.131922 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 21:39:24 crc kubenswrapper[4857]: I1201 21:39:24.317328 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 21:39:24 crc kubenswrapper[4857]: I1201 21:39:24.438928 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 21:39:24 crc kubenswrapper[4857]: I1201 21:39:24.556918 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 21:39:24 crc kubenswrapper[4857]: I1201 21:39:24.570563 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 21:39:24 crc kubenswrapper[4857]: I1201 21:39:24.907173 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 21:39:24 crc kubenswrapper[4857]: I1201 21:39:24.967966 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 21:39:25 crc kubenswrapper[4857]: I1201 21:39:25.127104 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 21:39:25 crc kubenswrapper[4857]: I1201 21:39:25.156229 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 21:39:25 crc kubenswrapper[4857]: I1201 21:39:25.307188 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 21:39:25 crc kubenswrapper[4857]: I1201 21:39:25.358563 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 21:39:25 crc kubenswrapper[4857]: I1201 21:39:25.415985 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 21:39:25 crc kubenswrapper[4857]: I1201 21:39:25.729188 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 21:39:25 crc kubenswrapper[4857]: I1201 21:39:25.766959 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 21:39:25 crc kubenswrapper[4857]: I1201 21:39:25.776593 4857 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 01 21:39:25 crc kubenswrapper[4857]: I1201 21:39:25.777105 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 01 21:39:25 crc kubenswrapper[4857]: I1201 21:39:25.789435 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 21:39:25 crc kubenswrapper[4857]: I1201 21:39:25.944929 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 21:39:25 crc kubenswrapper[4857]: I1201 21:39:25.962180 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.004423 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.069747 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.164778 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.201682 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.221535 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.223628 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.279254 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.471940 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.491342 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.518604 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.537316 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.719649 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.734120 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.746257 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.828874 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.939442 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.946797 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 21:39:26 crc kubenswrapper[4857]: I1201 21:39:26.948477 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.122468 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.281791 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.383611 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.405108 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.456173 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.573688 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.601923 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.624707 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.668477 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.698617 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.731515 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.741365 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.802220 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.848312 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 21:39:27 crc kubenswrapper[4857]: I1201 21:39:27.898436 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.012636 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.018221 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.088563 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.110022 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.124085 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.127446 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.256195 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.260843 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.284535 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.292003 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.415749 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.435279 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.469548 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.529189 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.649517 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.820612 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.838413 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.878600 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.924239 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 21:39:28 crc kubenswrapper[4857]: I1201 21:39:28.944364 4857 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.052156 4857 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.126496 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.154217 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.157945 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.179437 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.288216 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.315510 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.327093 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.353640 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.696508 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.814941 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.923131 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.937030 4857 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 21:39:29 crc kubenswrapper[4857]: I1201 21:39:29.967572 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.016988 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.074375 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.074534 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.074673 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.111389 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.123514 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.166625 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.168010 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.195190 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.219380 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.267246 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.268455 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.282557 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.316870 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.319579 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.370034 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.394532 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.400269 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.427022 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.452751 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.474843 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.493703 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.511847 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.519364 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.525760 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.598090 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.624189 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.648009 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.753652 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.781793 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.842256 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.861185 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.903404 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.934016 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.938309 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.939644 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 21:39:30 crc kubenswrapper[4857]: I1201 21:39:30.978118 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.003238 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.033458 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.072720 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.175575 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.210310 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.253456 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.263772 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.279715 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.281815 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.314256 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.375090 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.401260 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.527403 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.555571 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.577883 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.590827 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.641460 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.690318 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.721367 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.738674 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.870543 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 21:39:31 crc kubenswrapper[4857]: I1201 21:39:31.940929 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 21:39:32 crc kubenswrapper[4857]: I1201 21:39:32.036562 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 21:39:32 crc kubenswrapper[4857]: I1201 21:39:32.050987 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 21:39:32 crc kubenswrapper[4857]: I1201 21:39:32.062740 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 21:39:32 crc kubenswrapper[4857]: I1201 21:39:32.106322 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 21:39:32 crc kubenswrapper[4857]: I1201 21:39:32.286346 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 21:39:32 crc kubenswrapper[4857]: I1201 21:39:32.317093 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 21:39:32 crc kubenswrapper[4857]: I1201 21:39:32.413291 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 21:39:32 crc kubenswrapper[4857]: I1201 21:39:32.491543 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 21:39:32 crc kubenswrapper[4857]: I1201 21:39:32.593507 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 21:39:32 crc kubenswrapper[4857]: I1201 21:39:32.638516 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 21:39:32 crc kubenswrapper[4857]: I1201 21:39:32.664442 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.019516 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.128071 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.296258 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.364499 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.390831 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.407115 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.438582 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.439490 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.517510 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.538112 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.589318 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.608649 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.652276 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.661500 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.738606 4857 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.798351 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.868659 4857 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.931665 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 21:39:33 crc kubenswrapper[4857]: I1201 21:39:33.979354 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.006830 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.049491 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.084256 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.084312 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.094562 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.132216 4857 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.135362 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=44.135335299 podStartE2EDuration="44.135335299s" podCreationTimestamp="2025-12-01 21:38:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:39:12.063085096 +0000 UTC m=+270.553147413" watchObservedRunningTime="2025-12-01 21:39:34.135335299 +0000 UTC m=+292.625397656" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.142558 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.142652 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-84cc499644-zv4bt","openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 21:39:34 crc kubenswrapper[4857]: E1201 21:39:34.143078 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c350e136-f264-40ea-9b0e-9211dea68129" containerName="installer" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.143104 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c350e136-f264-40ea-9b0e-9211dea68129" containerName="installer" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.143316 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c350e136-f264-40ea-9b0e-9211dea68129" containerName="installer" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.143782 4857 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.143806 4857 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d25a23a8-5aa2-4324-8d55-e8000ac37979" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.144339 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.149780 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.149901 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.151613 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.151720 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.152017 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.152654 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.153347 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.153696 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.153911 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.154113 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.161667 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.161891 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.162299 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.168014 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.182000 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.189539 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.201203 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.201168795 podStartE2EDuration="22.201168795s" podCreationTimestamp="2025-12-01 21:39:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:39:34.199580758 +0000 UTC m=+292.689643115" watchObservedRunningTime="2025-12-01 21:39:34.201168795 +0000 UTC m=+292.691231122" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.201827 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.253876 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-user-template-login\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.253970 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-router-certs\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.253998 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.254028 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.254147 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-user-template-error\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.254240 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lzp9\" (UniqueName: \"kubernetes.io/projected/b7e45838-3601-4953-a0c9-d785b179562e-kube-api-access-5lzp9\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.254300 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7e45838-3601-4953-a0c9-d785b179562e-audit-policies\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.254349 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-session\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.254422 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-service-ca\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.254456 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.254541 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.254574 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7e45838-3601-4953-a0c9-d785b179562e-audit-dir\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.254618 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.254713 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.260592 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.331332 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355347 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-router-certs\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355389 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355419 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355436 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-user-template-error\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355454 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lzp9\" (UniqueName: \"kubernetes.io/projected/b7e45838-3601-4953-a0c9-d785b179562e-kube-api-access-5lzp9\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355472 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-session\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355490 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7e45838-3601-4953-a0c9-d785b179562e-audit-policies\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355512 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-service-ca\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355531 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355551 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355569 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7e45838-3601-4953-a0c9-d785b179562e-audit-dir\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355601 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355626 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.355657 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-user-template-login\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.356310 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7e45838-3601-4953-a0c9-d785b179562e-audit-dir\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.356853 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7e45838-3601-4953-a0c9-d785b179562e-audit-policies\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.356868 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-service-ca\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.357161 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.357224 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.358855 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.361365 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.361787 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-user-template-login\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.361789 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.362328 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-user-template-error\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.362906 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.363712 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-router-certs\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.364182 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.369566 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7e45838-3601-4953-a0c9-d785b179562e-v4-0-config-system-session\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.374784 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lzp9\" (UniqueName: \"kubernetes.io/projected/b7e45838-3601-4953-a0c9-d785b179562e-kube-api-access-5lzp9\") pod \"oauth-openshift-84cc499644-zv4bt\" (UID: \"b7e45838-3601-4953-a0c9-d785b179562e\") " pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.507800 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.541892 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.578483 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.644135 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.659798 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.704163 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.803653 4857 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.803867 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://db076f2683fbbb30d0b00304714478cf038f8a8272c27b3bec2b83b75bd8c928" gracePeriod=5 Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.819477 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-84cc499644-zv4bt"] Dec 01 21:39:34 crc kubenswrapper[4857]: W1201 21:39:34.826226 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7e45838_3601_4953_a0c9_d785b179562e.slice/crio-88a7774b8d7e6d7be1f115817bdf16ee5a2878c456efd078345c0163eb2648f4 WatchSource:0}: Error finding container 88a7774b8d7e6d7be1f115817bdf16ee5a2878c456efd078345c0163eb2648f4: Status 404 returned error can't find the container with id 88a7774b8d7e6d7be1f115817bdf16ee5a2878c456efd078345c0163eb2648f4 Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.857687 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 21:39:34 crc kubenswrapper[4857]: I1201 21:39:34.906651 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.051869 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.148874 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.177452 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.263871 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.335827 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" event={"ID":"b7e45838-3601-4953-a0c9-d785b179562e","Type":"ContainerStarted","Data":"7072c09f3b261420e361b826ed05858c1685fa0be684a0096433ccd226c5fdd4"} Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.335884 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" event={"ID":"b7e45838-3601-4953-a0c9-d785b179562e","Type":"ContainerStarted","Data":"88a7774b8d7e6d7be1f115817bdf16ee5a2878c456efd078345c0163eb2648f4"} Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.358774 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.370436 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" podStartSLOduration=75.370418154 podStartE2EDuration="1m15.370418154s" podCreationTimestamp="2025-12-01 21:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:39:35.367638513 +0000 UTC m=+293.857700830" watchObservedRunningTime="2025-12-01 21:39:35.370418154 +0000 UTC m=+293.860480471" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.570602 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.594696 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.616338 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.714463 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.775917 4857 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.776000 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.776089 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.776983 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"d600387d4b3fb1cfe43bd2b6cc1396aa79d3be00e6ce66722e0cbd9ea31e9a2b"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.777112 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://d600387d4b3fb1cfe43bd2b6cc1396aa79d3be00e6ce66722e0cbd9ea31e9a2b" gracePeriod=30 Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.820495 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.858593 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 21:39:35 crc kubenswrapper[4857]: I1201 21:39:35.967517 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.000358 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.026775 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.030635 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.200016 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.344213 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.350457 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-84cc499644-zv4bt" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.351489 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.451528 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.470326 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.559202 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.574513 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.630544 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.703758 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.810100 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 21:39:36 crc kubenswrapper[4857]: I1201 21:39:36.871173 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 21:39:37 crc kubenswrapper[4857]: I1201 21:39:37.045997 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 21:39:37 crc kubenswrapper[4857]: I1201 21:39:37.257602 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 21:39:37 crc kubenswrapper[4857]: I1201 21:39:37.282289 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 21:39:37 crc kubenswrapper[4857]: I1201 21:39:37.443560 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 21:39:37 crc kubenswrapper[4857]: I1201 21:39:37.509939 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 21:39:37 crc kubenswrapper[4857]: I1201 21:39:37.818940 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 21:39:38 crc kubenswrapper[4857]: I1201 21:39:38.021244 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 21:39:38 crc kubenswrapper[4857]: I1201 21:39:38.105937 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 21:39:38 crc kubenswrapper[4857]: I1201 21:39:38.968345 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.370870 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.371484 4857 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="db076f2683fbbb30d0b00304714478cf038f8a8272c27b3bec2b83b75bd8c928" exitCode=137 Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.371554 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7bbf941505864617f3256cfcbabeead7ab714914232940fd0c58d1ac74d019b" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.417806 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.417923 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613364 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613429 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613504 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613557 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613588 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613633 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613632 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613724 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613525 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613912 4857 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613929 4857 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613939 4857 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.613947 4857 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.627329 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:39:40 crc kubenswrapper[4857]: I1201 21:39:40.765646 4857 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 21:39:41 crc kubenswrapper[4857]: I1201 21:39:41.384421 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 21:39:41 crc kubenswrapper[4857]: I1201 21:39:41.687903 4857 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 01 21:39:41 crc kubenswrapper[4857]: I1201 21:39:41.846524 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 01 21:39:41 crc kubenswrapper[4857]: I1201 21:39:41.847779 4857 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 01 21:39:41 crc kubenswrapper[4857]: I1201 21:39:41.860939 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 21:39:41 crc kubenswrapper[4857]: I1201 21:39:41.860991 4857 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="a666e047-1e4d-45c3-8149-3614b16355c5" Dec 01 21:39:41 crc kubenswrapper[4857]: I1201 21:39:41.868712 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 21:39:41 crc kubenswrapper[4857]: I1201 21:39:41.868763 4857 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="a666e047-1e4d-45c3-8149-3614b16355c5" Dec 01 21:39:51 crc kubenswrapper[4857]: I1201 21:39:51.343908 4857 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mbzs2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 01 21:39:51 crc kubenswrapper[4857]: I1201 21:39:51.344116 4857 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mbzs2 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 01 21:39:51 crc kubenswrapper[4857]: I1201 21:39:51.344799 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" podUID="01903422-1e15-44cf-87f1-6037f6631592" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 01 21:39:51 crc kubenswrapper[4857]: I1201 21:39:51.344870 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" podUID="01903422-1e15-44cf-87f1-6037f6631592" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 01 21:39:51 crc kubenswrapper[4857]: I1201 21:39:51.490304 4857 generic.go:334] "Generic (PLEG): container finished" podID="01903422-1e15-44cf-87f1-6037f6631592" containerID="9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a" exitCode=0 Dec 01 21:39:51 crc kubenswrapper[4857]: I1201 21:39:51.490378 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" event={"ID":"01903422-1e15-44cf-87f1-6037f6631592","Type":"ContainerDied","Data":"9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a"} Dec 01 21:39:51 crc kubenswrapper[4857]: I1201 21:39:51.491161 4857 scope.go:117] "RemoveContainer" containerID="9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a" Dec 01 21:39:52 crc kubenswrapper[4857]: I1201 21:39:52.499753 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" event={"ID":"01903422-1e15-44cf-87f1-6037f6631592","Type":"ContainerStarted","Data":"d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef"} Dec 01 21:39:52 crc kubenswrapper[4857]: I1201 21:39:52.500776 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:39:52 crc kubenswrapper[4857]: I1201 21:39:52.506325 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:40:06 crc kubenswrapper[4857]: I1201 21:40:06.593723 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 01 21:40:06 crc kubenswrapper[4857]: I1201 21:40:06.598888 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 21:40:06 crc kubenswrapper[4857]: I1201 21:40:06.598983 4857 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="d600387d4b3fb1cfe43bd2b6cc1396aa79d3be00e6ce66722e0cbd9ea31e9a2b" exitCode=137 Dec 01 21:40:06 crc kubenswrapper[4857]: I1201 21:40:06.599062 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"d600387d4b3fb1cfe43bd2b6cc1396aa79d3be00e6ce66722e0cbd9ea31e9a2b"} Dec 01 21:40:06 crc kubenswrapper[4857]: I1201 21:40:06.599118 4857 scope.go:117] "RemoveContainer" containerID="f1a7eb8e5ad80a1f21309d1b7b5cc3c0943e3e781936497fad827417baa3adbc" Dec 01 21:40:07 crc kubenswrapper[4857]: I1201 21:40:07.606696 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 01 21:40:07 crc kubenswrapper[4857]: I1201 21:40:07.608989 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d728cd70698e5fef221428b192d8aa2980f1bebba659d480fdfd655aaec9a767"} Dec 01 21:40:07 crc kubenswrapper[4857]: I1201 21:40:07.862702 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:40:15 crc kubenswrapper[4857]: I1201 21:40:15.776056 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:40:15 crc kubenswrapper[4857]: I1201 21:40:15.788333 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:40:16 crc kubenswrapper[4857]: I1201 21:40:16.673391 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 21:40:27 crc kubenswrapper[4857]: I1201 21:40:27.551604 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr"] Dec 01 21:40:27 crc kubenswrapper[4857]: I1201 21:40:27.552299 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" podUID="5c55767b-b626-4e74-a0fe-23d71cf0e347" containerName="route-controller-manager" containerID="cri-o://f62e1960bd921e45a702c7c7ff0fb33a3ebcd178f24e33c43a196123c36f0814" gracePeriod=30 Dec 01 21:40:27 crc kubenswrapper[4857]: I1201 21:40:27.567357 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5wggv"] Dec 01 21:40:27 crc kubenswrapper[4857]: I1201 21:40:27.567558 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" podUID="71476481-b2d8-4a26-b8a8-66fa5ec7135e" containerName="controller-manager" containerID="cri-o://7b8c6257dd9710830ff496d98c9652e5d63a25ba837c0fed7aacc6d9fbf5be7f" gracePeriod=30 Dec 01 21:40:27 crc kubenswrapper[4857]: I1201 21:40:27.779489 4857 generic.go:334] "Generic (PLEG): container finished" podID="71476481-b2d8-4a26-b8a8-66fa5ec7135e" containerID="7b8c6257dd9710830ff496d98c9652e5d63a25ba837c0fed7aacc6d9fbf5be7f" exitCode=0 Dec 01 21:40:27 crc kubenswrapper[4857]: I1201 21:40:27.779562 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" event={"ID":"71476481-b2d8-4a26-b8a8-66fa5ec7135e","Type":"ContainerDied","Data":"7b8c6257dd9710830ff496d98c9652e5d63a25ba837c0fed7aacc6d9fbf5be7f"} Dec 01 21:40:27 crc kubenswrapper[4857]: I1201 21:40:27.783726 4857 generic.go:334] "Generic (PLEG): container finished" podID="5c55767b-b626-4e74-a0fe-23d71cf0e347" containerID="f62e1960bd921e45a702c7c7ff0fb33a3ebcd178f24e33c43a196123c36f0814" exitCode=0 Dec 01 21:40:27 crc kubenswrapper[4857]: I1201 21:40:27.783771 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" event={"ID":"5c55767b-b626-4e74-a0fe-23d71cf0e347","Type":"ContainerDied","Data":"f62e1960bd921e45a702c7c7ff0fb33a3ebcd178f24e33c43a196123c36f0814"} Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.056286 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.063294 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.151251 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c55767b-b626-4e74-a0fe-23d71cf0e347-config\") pod \"5c55767b-b626-4e74-a0fe-23d71cf0e347\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.151327 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m86dd\" (UniqueName: \"kubernetes.io/projected/5c55767b-b626-4e74-a0fe-23d71cf0e347-kube-api-access-m86dd\") pod \"5c55767b-b626-4e74-a0fe-23d71cf0e347\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.151414 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c55767b-b626-4e74-a0fe-23d71cf0e347-serving-cert\") pod \"5c55767b-b626-4e74-a0fe-23d71cf0e347\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.151520 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c55767b-b626-4e74-a0fe-23d71cf0e347-client-ca\") pod \"5c55767b-b626-4e74-a0fe-23d71cf0e347\" (UID: \"5c55767b-b626-4e74-a0fe-23d71cf0e347\") " Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.152235 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c55767b-b626-4e74-a0fe-23d71cf0e347-config" (OuterVolumeSpecName: "config") pod "5c55767b-b626-4e74-a0fe-23d71cf0e347" (UID: "5c55767b-b626-4e74-a0fe-23d71cf0e347"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.152379 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c55767b-b626-4e74-a0fe-23d71cf0e347-client-ca" (OuterVolumeSpecName: "client-ca") pod "5c55767b-b626-4e74-a0fe-23d71cf0e347" (UID: "5c55767b-b626-4e74-a0fe-23d71cf0e347"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.158298 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c55767b-b626-4e74-a0fe-23d71cf0e347-kube-api-access-m86dd" (OuterVolumeSpecName: "kube-api-access-m86dd") pod "5c55767b-b626-4e74-a0fe-23d71cf0e347" (UID: "5c55767b-b626-4e74-a0fe-23d71cf0e347"). InnerVolumeSpecName "kube-api-access-m86dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.162188 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c55767b-b626-4e74-a0fe-23d71cf0e347-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5c55767b-b626-4e74-a0fe-23d71cf0e347" (UID: "5c55767b-b626-4e74-a0fe-23d71cf0e347"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.253238 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-client-ca\") pod \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.253365 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71476481-b2d8-4a26-b8a8-66fa5ec7135e-serving-cert\") pod \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.253407 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlljj\" (UniqueName: \"kubernetes.io/projected/71476481-b2d8-4a26-b8a8-66fa5ec7135e-kube-api-access-jlljj\") pod \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.253442 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-config\") pod \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.253468 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-proxy-ca-bundles\") pod \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\" (UID: \"71476481-b2d8-4a26-b8a8-66fa5ec7135e\") " Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.253839 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c55767b-b626-4e74-a0fe-23d71cf0e347-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.253871 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m86dd\" (UniqueName: \"kubernetes.io/projected/5c55767b-b626-4e74-a0fe-23d71cf0e347-kube-api-access-m86dd\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.253891 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c55767b-b626-4e74-a0fe-23d71cf0e347-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.253906 4857 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c55767b-b626-4e74-a0fe-23d71cf0e347-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.254875 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-config" (OuterVolumeSpecName: "config") pod "71476481-b2d8-4a26-b8a8-66fa5ec7135e" (UID: "71476481-b2d8-4a26-b8a8-66fa5ec7135e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.255004 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-client-ca" (OuterVolumeSpecName: "client-ca") pod "71476481-b2d8-4a26-b8a8-66fa5ec7135e" (UID: "71476481-b2d8-4a26-b8a8-66fa5ec7135e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.255104 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "71476481-b2d8-4a26-b8a8-66fa5ec7135e" (UID: "71476481-b2d8-4a26-b8a8-66fa5ec7135e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.257285 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71476481-b2d8-4a26-b8a8-66fa5ec7135e-kube-api-access-jlljj" (OuterVolumeSpecName: "kube-api-access-jlljj") pod "71476481-b2d8-4a26-b8a8-66fa5ec7135e" (UID: "71476481-b2d8-4a26-b8a8-66fa5ec7135e"). InnerVolumeSpecName "kube-api-access-jlljj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.259286 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71476481-b2d8-4a26-b8a8-66fa5ec7135e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "71476481-b2d8-4a26-b8a8-66fa5ec7135e" (UID: "71476481-b2d8-4a26-b8a8-66fa5ec7135e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.355186 4857 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.355228 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71476481-b2d8-4a26-b8a8-66fa5ec7135e-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.355245 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlljj\" (UniqueName: \"kubernetes.io/projected/71476481-b2d8-4a26-b8a8-66fa5ec7135e-kube-api-access-jlljj\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.355261 4857 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.355274 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71476481-b2d8-4a26-b8a8-66fa5ec7135e-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.791745 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" event={"ID":"5c55767b-b626-4e74-a0fe-23d71cf0e347","Type":"ContainerDied","Data":"e3efe52763acd8d35ae2e0299a2af5806dfebd75e6b4797580ee885b8f138b44"} Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.791807 4857 scope.go:117] "RemoveContainer" containerID="f62e1960bd921e45a702c7c7ff0fb33a3ebcd178f24e33c43a196123c36f0814" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.793419 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.795689 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" event={"ID":"71476481-b2d8-4a26-b8a8-66fa5ec7135e","Type":"ContainerDied","Data":"c14c303356ee459aa88762763e0742a5185b1d073917334d9529fb24976065cb"} Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.795839 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5wggv" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.820550 4857 scope.go:117] "RemoveContainer" containerID="7b8c6257dd9710830ff496d98c9652e5d63a25ba837c0fed7aacc6d9fbf5be7f" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.833284 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr"] Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.843687 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8rjjr"] Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.851904 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5wggv"] Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.860162 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5wggv"] Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.985757 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66f9447cbb-bx7gq"] Dec 01 21:40:28 crc kubenswrapper[4857]: E1201 21:40:28.985992 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c55767b-b626-4e74-a0fe-23d71cf0e347" containerName="route-controller-manager" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.986004 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c55767b-b626-4e74-a0fe-23d71cf0e347" containerName="route-controller-manager" Dec 01 21:40:28 crc kubenswrapper[4857]: E1201 21:40:28.986015 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71476481-b2d8-4a26-b8a8-66fa5ec7135e" containerName="controller-manager" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.986021 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="71476481-b2d8-4a26-b8a8-66fa5ec7135e" containerName="controller-manager" Dec 01 21:40:28 crc kubenswrapper[4857]: E1201 21:40:28.986031 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.986050 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.986144 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c55767b-b626-4e74-a0fe-23d71cf0e347" containerName="route-controller-manager" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.986158 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="71476481-b2d8-4a26-b8a8-66fa5ec7135e" containerName="controller-manager" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.986166 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.986562 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.989772 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8"] Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.990573 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.991896 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.992514 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.992948 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.993509 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.993805 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.993998 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.994182 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.994467 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.994506 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.994814 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 21:40:28 crc kubenswrapper[4857]: I1201 21:40:28.995088 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.004355 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.012350 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8"] Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.027106 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.028093 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66f9447cbb-bx7gq"] Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.165411 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-proxy-ca-bundles\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.165460 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c95f8562-7dcc-4f89-b839-d46b23394b10-config\") pod \"route-controller-manager-5b78479656-n9rh8\" (UID: \"c95f8562-7dcc-4f89-b839-d46b23394b10\") " pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.165478 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c95f8562-7dcc-4f89-b839-d46b23394b10-serving-cert\") pod \"route-controller-manager-5b78479656-n9rh8\" (UID: \"c95f8562-7dcc-4f89-b839-d46b23394b10\") " pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.165501 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-client-ca\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.165514 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-serving-cert\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.165544 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-config\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.165565 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c95f8562-7dcc-4f89-b839-d46b23394b10-client-ca\") pod \"route-controller-manager-5b78479656-n9rh8\" (UID: \"c95f8562-7dcc-4f89-b839-d46b23394b10\") " pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.165581 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pmgc\" (UniqueName: \"kubernetes.io/projected/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-kube-api-access-7pmgc\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.165757 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r67vm\" (UniqueName: \"kubernetes.io/projected/c95f8562-7dcc-4f89-b839-d46b23394b10-kube-api-access-r67vm\") pod \"route-controller-manager-5b78479656-n9rh8\" (UID: \"c95f8562-7dcc-4f89-b839-d46b23394b10\") " pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.267526 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-config\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.267608 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c95f8562-7dcc-4f89-b839-d46b23394b10-client-ca\") pod \"route-controller-manager-5b78479656-n9rh8\" (UID: \"c95f8562-7dcc-4f89-b839-d46b23394b10\") " pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.267643 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pmgc\" (UniqueName: \"kubernetes.io/projected/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-kube-api-access-7pmgc\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.267689 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r67vm\" (UniqueName: \"kubernetes.io/projected/c95f8562-7dcc-4f89-b839-d46b23394b10-kube-api-access-r67vm\") pod \"route-controller-manager-5b78479656-n9rh8\" (UID: \"c95f8562-7dcc-4f89-b839-d46b23394b10\") " pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.267786 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-proxy-ca-bundles\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.267831 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c95f8562-7dcc-4f89-b839-d46b23394b10-config\") pod \"route-controller-manager-5b78479656-n9rh8\" (UID: \"c95f8562-7dcc-4f89-b839-d46b23394b10\") " pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.267860 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c95f8562-7dcc-4f89-b839-d46b23394b10-serving-cert\") pod \"route-controller-manager-5b78479656-n9rh8\" (UID: \"c95f8562-7dcc-4f89-b839-d46b23394b10\") " pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.267900 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-client-ca\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.267927 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-serving-cert\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.269935 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-proxy-ca-bundles\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.270925 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c95f8562-7dcc-4f89-b839-d46b23394b10-client-ca\") pod \"route-controller-manager-5b78479656-n9rh8\" (UID: \"c95f8562-7dcc-4f89-b839-d46b23394b10\") " pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.272624 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-serving-cert\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.272735 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-config\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.273844 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-client-ca\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.273879 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c95f8562-7dcc-4f89-b839-d46b23394b10-config\") pod \"route-controller-manager-5b78479656-n9rh8\" (UID: \"c95f8562-7dcc-4f89-b839-d46b23394b10\") " pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.282936 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c95f8562-7dcc-4f89-b839-d46b23394b10-serving-cert\") pod \"route-controller-manager-5b78479656-n9rh8\" (UID: \"c95f8562-7dcc-4f89-b839-d46b23394b10\") " pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.295324 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r67vm\" (UniqueName: \"kubernetes.io/projected/c95f8562-7dcc-4f89-b839-d46b23394b10-kube-api-access-r67vm\") pod \"route-controller-manager-5b78479656-n9rh8\" (UID: \"c95f8562-7dcc-4f89-b839-d46b23394b10\") " pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.295976 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pmgc\" (UniqueName: \"kubernetes.io/projected/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-kube-api-access-7pmgc\") pod \"controller-manager-66f9447cbb-bx7gq\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.313724 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.333106 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.649362 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8"] Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.803787 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" event={"ID":"c95f8562-7dcc-4f89-b839-d46b23394b10","Type":"ContainerStarted","Data":"193820566c591c6a29ef8650dae57cf3359162095edf6d7e910ba69a914491f3"} Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.804305 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.804335 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" event={"ID":"c95f8562-7dcc-4f89-b839-d46b23394b10","Type":"ContainerStarted","Data":"f740e4a7ba712d5f86b6d687778e1254ae7ca2c5525745524674f4332f294913"} Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.806320 4857 patch_prober.go:28] interesting pod/route-controller-manager-5b78479656-n9rh8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.806451 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" podUID="c95f8562-7dcc-4f89-b839-d46b23394b10" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.807780 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66f9447cbb-bx7gq"] Dec 01 21:40:29 crc kubenswrapper[4857]: W1201 21:40:29.815387 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc19c1f73_f13a_4442_bdb4_e29f2cfa98b2.slice/crio-00b6a1cdcf1ecd04eb28d068787efb563b706aafe6953421fcf7a8700442c872 WatchSource:0}: Error finding container 00b6a1cdcf1ecd04eb28d068787efb563b706aafe6953421fcf7a8700442c872: Status 404 returned error can't find the container with id 00b6a1cdcf1ecd04eb28d068787efb563b706aafe6953421fcf7a8700442c872 Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.824385 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" podStartSLOduration=2.824363444 podStartE2EDuration="2.824363444s" podCreationTimestamp="2025-12-01 21:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:40:29.821868282 +0000 UTC m=+348.311930599" watchObservedRunningTime="2025-12-01 21:40:29.824363444 +0000 UTC m=+348.314425771" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.842406 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c55767b-b626-4e74-a0fe-23d71cf0e347" path="/var/lib/kubelet/pods/5c55767b-b626-4e74-a0fe-23d71cf0e347/volumes" Dec 01 21:40:29 crc kubenswrapper[4857]: I1201 21:40:29.843315 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71476481-b2d8-4a26-b8a8-66fa5ec7135e" path="/var/lib/kubelet/pods/71476481-b2d8-4a26-b8a8-66fa5ec7135e/volumes" Dec 01 21:40:30 crc kubenswrapper[4857]: I1201 21:40:30.832856 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" event={"ID":"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2","Type":"ContainerStarted","Data":"fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab"} Dec 01 21:40:30 crc kubenswrapper[4857]: I1201 21:40:30.834454 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" event={"ID":"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2","Type":"ContainerStarted","Data":"00b6a1cdcf1ecd04eb28d068787efb563b706aafe6953421fcf7a8700442c872"} Dec 01 21:40:30 crc kubenswrapper[4857]: I1201 21:40:30.843999 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5b78479656-n9rh8" Dec 01 21:40:30 crc kubenswrapper[4857]: I1201 21:40:30.864159 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" podStartSLOduration=3.864136366 podStartE2EDuration="3.864136366s" podCreationTimestamp="2025-12-01 21:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:40:30.858781129 +0000 UTC m=+349.348843456" watchObservedRunningTime="2025-12-01 21:40:30.864136366 +0000 UTC m=+349.354198683" Dec 01 21:40:31 crc kubenswrapper[4857]: I1201 21:40:31.845606 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:31 crc kubenswrapper[4857]: I1201 21:40:31.845890 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:36 crc kubenswrapper[4857]: I1201 21:40:36.851970 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66f9447cbb-bx7gq"] Dec 01 21:40:36 crc kubenswrapper[4857]: I1201 21:40:36.852638 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" podUID="c19c1f73-f13a-4442-bdb4-e29f2cfa98b2" containerName="controller-manager" containerID="cri-o://fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab" gracePeriod=30 Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.386987 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.416679 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-config\") pod \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.416725 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-serving-cert\") pod \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.416748 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pmgc\" (UniqueName: \"kubernetes.io/projected/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-kube-api-access-7pmgc\") pod \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.416786 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-proxy-ca-bundles\") pod \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.416828 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-client-ca\") pod \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\" (UID: \"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2\") " Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.417791 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c19c1f73-f13a-4442-bdb4-e29f2cfa98b2" (UID: "c19c1f73-f13a-4442-bdb4-e29f2cfa98b2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.417826 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-client-ca" (OuterVolumeSpecName: "client-ca") pod "c19c1f73-f13a-4442-bdb4-e29f2cfa98b2" (UID: "c19c1f73-f13a-4442-bdb4-e29f2cfa98b2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.417840 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-config" (OuterVolumeSpecName: "config") pod "c19c1f73-f13a-4442-bdb4-e29f2cfa98b2" (UID: "c19c1f73-f13a-4442-bdb4-e29f2cfa98b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.432684 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-kube-api-access-7pmgc" (OuterVolumeSpecName: "kube-api-access-7pmgc") pod "c19c1f73-f13a-4442-bdb4-e29f2cfa98b2" (UID: "c19c1f73-f13a-4442-bdb4-e29f2cfa98b2"). InnerVolumeSpecName "kube-api-access-7pmgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.436094 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c19c1f73-f13a-4442-bdb4-e29f2cfa98b2" (UID: "c19c1f73-f13a-4442-bdb4-e29f2cfa98b2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.518071 4857 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.518513 4857 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.518534 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.518554 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.518573 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pmgc\" (UniqueName: \"kubernetes.io/projected/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2-kube-api-access-7pmgc\") on node \"crc\" DevicePath \"\"" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.881171 4857 generic.go:334] "Generic (PLEG): container finished" podID="c19c1f73-f13a-4442-bdb4-e29f2cfa98b2" containerID="fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab" exitCode=0 Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.881307 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" event={"ID":"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2","Type":"ContainerDied","Data":"fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab"} Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.881323 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.881366 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f9447cbb-bx7gq" event={"ID":"c19c1f73-f13a-4442-bdb4-e29f2cfa98b2","Type":"ContainerDied","Data":"00b6a1cdcf1ecd04eb28d068787efb563b706aafe6953421fcf7a8700442c872"} Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.881396 4857 scope.go:117] "RemoveContainer" containerID="fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.913678 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66f9447cbb-bx7gq"] Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.917291 4857 scope.go:117] "RemoveContainer" containerID="fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab" Dec 01 21:40:37 crc kubenswrapper[4857]: E1201 21:40:37.917907 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab\": container with ID starting with fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab not found: ID does not exist" containerID="fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.917959 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab"} err="failed to get container status \"fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab\": rpc error: code = NotFound desc = could not find container \"fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab\": container with ID starting with fbcf5520ee0f738f331207f6c736293485793560e325aa3561cc5f8618310cab not found: ID does not exist" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.921968 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-66f9447cbb-bx7gq"] Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.991561 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8"] Dec 01 21:40:37 crc kubenswrapper[4857]: E1201 21:40:37.991771 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19c1f73-f13a-4442-bdb4-e29f2cfa98b2" containerName="controller-manager" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.991783 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19c1f73-f13a-4442-bdb4-e29f2cfa98b2" containerName="controller-manager" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.991870 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c19c1f73-f13a-4442-bdb4-e29f2cfa98b2" containerName="controller-manager" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.992305 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.998120 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.998287 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.998491 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.998129 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.999319 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 21:40:37 crc kubenswrapper[4857]: I1201 21:40:37.999500 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.025666 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.025749 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d70587a-0db9-467c-b086-b2585d1bab75-proxy-ca-bundles\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.025804 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d70587a-0db9-467c-b086-b2585d1bab75-config\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.025832 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d70587a-0db9-467c-b086-b2585d1bab75-client-ca\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.025899 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d70587a-0db9-467c-b086-b2585d1bab75-serving-cert\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.025929 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkzlv\" (UniqueName: \"kubernetes.io/projected/2d70587a-0db9-467c-b086-b2585d1bab75-kube-api-access-mkzlv\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.038295 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8"] Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.127537 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d70587a-0db9-467c-b086-b2585d1bab75-config\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.127605 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d70587a-0db9-467c-b086-b2585d1bab75-client-ca\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.127888 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d70587a-0db9-467c-b086-b2585d1bab75-serving-cert\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.128110 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkzlv\" (UniqueName: \"kubernetes.io/projected/2d70587a-0db9-467c-b086-b2585d1bab75-kube-api-access-mkzlv\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.128777 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d70587a-0db9-467c-b086-b2585d1bab75-proxy-ca-bundles\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.130759 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d70587a-0db9-467c-b086-b2585d1bab75-config\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.131129 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d70587a-0db9-467c-b086-b2585d1bab75-client-ca\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.132185 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d70587a-0db9-467c-b086-b2585d1bab75-proxy-ca-bundles\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.134329 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d70587a-0db9-467c-b086-b2585d1bab75-serving-cert\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.154663 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkzlv\" (UniqueName: \"kubernetes.io/projected/2d70587a-0db9-467c-b086-b2585d1bab75-kube-api-access-mkzlv\") pod \"controller-manager-7c4cc668b4-xdbn8\" (UID: \"2d70587a-0db9-467c-b086-b2585d1bab75\") " pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.335400 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.865229 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8"] Dec 01 21:40:38 crc kubenswrapper[4857]: I1201 21:40:38.890518 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" event={"ID":"2d70587a-0db9-467c-b086-b2585d1bab75","Type":"ContainerStarted","Data":"07ef06281091af8b2fa103bfdef18537f6fc6f4382d60fb52a491b72da3a3b08"} Dec 01 21:40:39 crc kubenswrapper[4857]: I1201 21:40:39.841058 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c19c1f73-f13a-4442-bdb4-e29f2cfa98b2" path="/var/lib/kubelet/pods/c19c1f73-f13a-4442-bdb4-e29f2cfa98b2/volumes" Dec 01 21:40:39 crc kubenswrapper[4857]: I1201 21:40:39.900396 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" event={"ID":"2d70587a-0db9-467c-b086-b2585d1bab75","Type":"ContainerStarted","Data":"4066efea16b6be11c7f1c3666fa0ceae0d6818daecc71100a0d08e534f859287"} Dec 01 21:40:39 crc kubenswrapper[4857]: I1201 21:40:39.901099 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:39 crc kubenswrapper[4857]: I1201 21:40:39.906135 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" Dec 01 21:40:39 crc kubenswrapper[4857]: I1201 21:40:39.927471 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c4cc668b4-xdbn8" podStartSLOduration=3.927440646 podStartE2EDuration="3.927440646s" podCreationTimestamp="2025-12-01 21:40:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:40:39.92552627 +0000 UTC m=+358.415588617" watchObservedRunningTime="2025-12-01 21:40:39.927440646 +0000 UTC m=+358.417503003" Dec 01 21:40:57 crc kubenswrapper[4857]: I1201 21:40:57.806691 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:40:57 crc kubenswrapper[4857]: I1201 21:40:57.807322 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.361878 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4s5mw"] Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.363162 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4s5mw" podUID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" containerName="registry-server" containerID="cri-o://6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f" gracePeriod=30 Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.374199 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k9szd"] Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.374559 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k9szd" podUID="8049f37f-5448-4796-b384-691649bd559c" containerName="registry-server" containerID="cri-o://511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5" gracePeriod=30 Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.397508 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mbzs2"] Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.397860 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" podUID="01903422-1e15-44cf-87f1-6037f6631592" containerName="marketplace-operator" containerID="cri-o://d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef" gracePeriod=30 Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.407017 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-thg27"] Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.411455 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-thg27" podUID="2235e98c-f4f9-40ae-8f47-788c336e5893" containerName="registry-server" containerID="cri-o://76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56" gracePeriod=30 Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.433473 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zrwmc"] Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.434056 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zrwmc" podUID="226b6f72-5676-4dc8-9536-d99da01652c8" containerName="registry-server" containerID="cri-o://f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf" gracePeriod=30 Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.476260 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d5gjm"] Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.478822 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.482881 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d5gjm"] Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.620595 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwkbc\" (UniqueName: \"kubernetes.io/projected/9fe4a11e-f252-4f61-a8f4-d513b2b40643-kube-api-access-fwkbc\") pod \"marketplace-operator-79b997595-d5gjm\" (UID: \"9fe4a11e-f252-4f61-a8f4-d513b2b40643\") " pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.620663 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9fe4a11e-f252-4f61-a8f4-d513b2b40643-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d5gjm\" (UID: \"9fe4a11e-f252-4f61-a8f4-d513b2b40643\") " pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.620718 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9fe4a11e-f252-4f61-a8f4-d513b2b40643-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d5gjm\" (UID: \"9fe4a11e-f252-4f61-a8f4-d513b2b40643\") " pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.722380 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwkbc\" (UniqueName: \"kubernetes.io/projected/9fe4a11e-f252-4f61-a8f4-d513b2b40643-kube-api-access-fwkbc\") pod \"marketplace-operator-79b997595-d5gjm\" (UID: \"9fe4a11e-f252-4f61-a8f4-d513b2b40643\") " pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.722456 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9fe4a11e-f252-4f61-a8f4-d513b2b40643-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d5gjm\" (UID: \"9fe4a11e-f252-4f61-a8f4-d513b2b40643\") " pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.722539 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9fe4a11e-f252-4f61-a8f4-d513b2b40643-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d5gjm\" (UID: \"9fe4a11e-f252-4f61-a8f4-d513b2b40643\") " pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.725887 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9fe4a11e-f252-4f61-a8f4-d513b2b40643-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d5gjm\" (UID: \"9fe4a11e-f252-4f61-a8f4-d513b2b40643\") " pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.732407 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9fe4a11e-f252-4f61-a8f4-d513b2b40643-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d5gjm\" (UID: \"9fe4a11e-f252-4f61-a8f4-d513b2b40643\") " pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.743427 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwkbc\" (UniqueName: \"kubernetes.io/projected/9fe4a11e-f252-4f61-a8f4-d513b2b40643-kube-api-access-fwkbc\") pod \"marketplace-operator-79b997595-d5gjm\" (UID: \"9fe4a11e-f252-4f61-a8f4-d513b2b40643\") " pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.915936 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.922941 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:41:18 crc kubenswrapper[4857]: I1201 21:41:18.925238 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.001844 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.025919 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.028852 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phc6k\" (UniqueName: \"kubernetes.io/projected/8049f37f-5448-4796-b384-691649bd559c-kube-api-access-phc6k\") pod \"8049f37f-5448-4796-b384-691649bd559c\" (UID: \"8049f37f-5448-4796-b384-691649bd559c\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.029017 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-catalog-content\") pod \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\" (UID: \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.029059 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8049f37f-5448-4796-b384-691649bd559c-catalog-content\") pod \"8049f37f-5448-4796-b384-691649bd559c\" (UID: \"8049f37f-5448-4796-b384-691649bd559c\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.029136 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szw7m\" (UniqueName: \"kubernetes.io/projected/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-kube-api-access-szw7m\") pod \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\" (UID: \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.029160 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-utilities\") pod \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\" (UID: \"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.029187 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8049f37f-5448-4796-b384-691649bd559c-utilities\") pod \"8049f37f-5448-4796-b384-691649bd559c\" (UID: \"8049f37f-5448-4796-b384-691649bd559c\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.031195 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-utilities" (OuterVolumeSpecName: "utilities") pod "2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" (UID: "2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.036320 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-kube-api-access-szw7m" (OuterVolumeSpecName: "kube-api-access-szw7m") pod "2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" (UID: "2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a"). InnerVolumeSpecName "kube-api-access-szw7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.036803 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8049f37f-5448-4796-b384-691649bd559c-kube-api-access-phc6k" (OuterVolumeSpecName: "kube-api-access-phc6k") pod "8049f37f-5448-4796-b384-691649bd559c" (UID: "8049f37f-5448-4796-b384-691649bd559c"). InnerVolumeSpecName "kube-api-access-phc6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.037347 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8049f37f-5448-4796-b384-691649bd559c-utilities" (OuterVolumeSpecName: "utilities") pod "8049f37f-5448-4796-b384-691649bd559c" (UID: "8049f37f-5448-4796-b384-691649bd559c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.041267 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.104649 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" (UID: "2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.104816 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8049f37f-5448-4796-b384-691649bd559c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8049f37f-5448-4796-b384-691649bd559c" (UID: "8049f37f-5448-4796-b384-691649bd559c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.130742 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/226b6f72-5676-4dc8-9536-d99da01652c8-utilities\") pod \"226b6f72-5676-4dc8-9536-d99da01652c8\" (UID: \"226b6f72-5676-4dc8-9536-d99da01652c8\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.130777 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dd2bq\" (UniqueName: \"kubernetes.io/projected/226b6f72-5676-4dc8-9536-d99da01652c8-kube-api-access-dd2bq\") pod \"226b6f72-5676-4dc8-9536-d99da01652c8\" (UID: \"226b6f72-5676-4dc8-9536-d99da01652c8\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.130798 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/226b6f72-5676-4dc8-9536-d99da01652c8-catalog-content\") pod \"226b6f72-5676-4dc8-9536-d99da01652c8\" (UID: \"226b6f72-5676-4dc8-9536-d99da01652c8\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.130826 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fvpq\" (UniqueName: \"kubernetes.io/projected/2235e98c-f4f9-40ae-8f47-788c336e5893-kube-api-access-9fvpq\") pod \"2235e98c-f4f9-40ae-8f47-788c336e5893\" (UID: \"2235e98c-f4f9-40ae-8f47-788c336e5893\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131482 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/226b6f72-5676-4dc8-9536-d99da01652c8-utilities" (OuterVolumeSpecName: "utilities") pod "226b6f72-5676-4dc8-9536-d99da01652c8" (UID: "226b6f72-5676-4dc8-9536-d99da01652c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131567 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2235e98c-f4f9-40ae-8f47-788c336e5893-catalog-content\") pod \"2235e98c-f4f9-40ae-8f47-788c336e5893\" (UID: \"2235e98c-f4f9-40ae-8f47-788c336e5893\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131595 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01903422-1e15-44cf-87f1-6037f6631592-marketplace-trusted-ca\") pod \"01903422-1e15-44cf-87f1-6037f6631592\" (UID: \"01903422-1e15-44cf-87f1-6037f6631592\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131646 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2235e98c-f4f9-40ae-8f47-788c336e5893-utilities\") pod \"2235e98c-f4f9-40ae-8f47-788c336e5893\" (UID: \"2235e98c-f4f9-40ae-8f47-788c336e5893\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131663 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ng9h\" (UniqueName: \"kubernetes.io/projected/01903422-1e15-44cf-87f1-6037f6631592-kube-api-access-5ng9h\") pod \"01903422-1e15-44cf-87f1-6037f6631592\" (UID: \"01903422-1e15-44cf-87f1-6037f6631592\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131704 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/01903422-1e15-44cf-87f1-6037f6631592-marketplace-operator-metrics\") pod \"01903422-1e15-44cf-87f1-6037f6631592\" (UID: \"01903422-1e15-44cf-87f1-6037f6631592\") " Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131900 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szw7m\" (UniqueName: \"kubernetes.io/projected/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-kube-api-access-szw7m\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131913 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131921 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8049f37f-5448-4796-b384-691649bd559c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131930 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phc6k\" (UniqueName: \"kubernetes.io/projected/8049f37f-5448-4796-b384-691649bd559c-kube-api-access-phc6k\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131938 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/226b6f72-5676-4dc8-9536-d99da01652c8-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131946 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.131956 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8049f37f-5448-4796-b384-691649bd559c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.132419 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01903422-1e15-44cf-87f1-6037f6631592-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "01903422-1e15-44cf-87f1-6037f6631592" (UID: "01903422-1e15-44cf-87f1-6037f6631592"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.136615 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2235e98c-f4f9-40ae-8f47-788c336e5893-kube-api-access-9fvpq" (OuterVolumeSpecName: "kube-api-access-9fvpq") pod "2235e98c-f4f9-40ae-8f47-788c336e5893" (UID: "2235e98c-f4f9-40ae-8f47-788c336e5893"). InnerVolumeSpecName "kube-api-access-9fvpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.137266 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/226b6f72-5676-4dc8-9536-d99da01652c8-kube-api-access-dd2bq" (OuterVolumeSpecName: "kube-api-access-dd2bq") pod "226b6f72-5676-4dc8-9536-d99da01652c8" (UID: "226b6f72-5676-4dc8-9536-d99da01652c8"). InnerVolumeSpecName "kube-api-access-dd2bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.137470 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01903422-1e15-44cf-87f1-6037f6631592-kube-api-access-5ng9h" (OuterVolumeSpecName: "kube-api-access-5ng9h") pod "01903422-1e15-44cf-87f1-6037f6631592" (UID: "01903422-1e15-44cf-87f1-6037f6631592"). InnerVolumeSpecName "kube-api-access-5ng9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.137861 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2235e98c-f4f9-40ae-8f47-788c336e5893-utilities" (OuterVolumeSpecName: "utilities") pod "2235e98c-f4f9-40ae-8f47-788c336e5893" (UID: "2235e98c-f4f9-40ae-8f47-788c336e5893"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.141862 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01903422-1e15-44cf-87f1-6037f6631592-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "01903422-1e15-44cf-87f1-6037f6631592" (UID: "01903422-1e15-44cf-87f1-6037f6631592"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.153940 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2235e98c-f4f9-40ae-8f47-788c336e5893-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2235e98c-f4f9-40ae-8f47-788c336e5893" (UID: "2235e98c-f4f9-40ae-8f47-788c336e5893"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.180652 4857 generic.go:334] "Generic (PLEG): container finished" podID="8049f37f-5448-4796-b384-691649bd559c" containerID="511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5" exitCode=0 Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.180745 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9szd" event={"ID":"8049f37f-5448-4796-b384-691649bd559c","Type":"ContainerDied","Data":"511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5"} Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.180778 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9szd" event={"ID":"8049f37f-5448-4796-b384-691649bd559c","Type":"ContainerDied","Data":"67667efc4b7857e24fb75b7f672ccce6c0167a8ec2385a0f5e1e19546291f49f"} Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.180795 4857 scope.go:117] "RemoveContainer" containerID="511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.180913 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k9szd" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.184607 4857 generic.go:334] "Generic (PLEG): container finished" podID="2235e98c-f4f9-40ae-8f47-788c336e5893" containerID="76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56" exitCode=0 Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.184648 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thg27" event={"ID":"2235e98c-f4f9-40ae-8f47-788c336e5893","Type":"ContainerDied","Data":"76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56"} Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.184770 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thg27" event={"ID":"2235e98c-f4f9-40ae-8f47-788c336e5893","Type":"ContainerDied","Data":"b1b516c485e262175113e0f6e35038996f6b7453acff478d58a2fd4beefec640"} Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.184913 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thg27" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.187433 4857 generic.go:334] "Generic (PLEG): container finished" podID="01903422-1e15-44cf-87f1-6037f6631592" containerID="d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef" exitCode=0 Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.187520 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.187528 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" event={"ID":"01903422-1e15-44cf-87f1-6037f6631592","Type":"ContainerDied","Data":"d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef"} Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.187793 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mbzs2" event={"ID":"01903422-1e15-44cf-87f1-6037f6631592","Type":"ContainerDied","Data":"e73b46dff6eae12c378780cdbf27e9ebb112abe51ac617a3b1d30b5516d06391"} Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.189872 4857 generic.go:334] "Generic (PLEG): container finished" podID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" containerID="6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f" exitCode=0 Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.189914 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4s5mw" event={"ID":"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a","Type":"ContainerDied","Data":"6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f"} Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.189934 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4s5mw" event={"ID":"2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a","Type":"ContainerDied","Data":"7f24f22b6e81f45b5b03ed5d918ad908d387f243bb946e159a3e2abd9333cdb8"} Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.189993 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4s5mw" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.191713 4857 generic.go:334] "Generic (PLEG): container finished" podID="226b6f72-5676-4dc8-9536-d99da01652c8" containerID="f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf" exitCode=0 Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.191732 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrwmc" event={"ID":"226b6f72-5676-4dc8-9536-d99da01652c8","Type":"ContainerDied","Data":"f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf"} Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.191748 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrwmc" event={"ID":"226b6f72-5676-4dc8-9536-d99da01652c8","Type":"ContainerDied","Data":"8446b2a1eeb243d5baeb89e1ed393cc9469b8fc513c38444416a9a160fef3720"} Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.191792 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zrwmc" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.212874 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k9szd"] Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.216172 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k9szd"] Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.222182 4857 scope.go:117] "RemoveContainer" containerID="4330ab6c16a16ed62e8ea9df65b84cbf1e29ab7f3560c278f9884f8ab04914a2" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.232986 4857 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/01903422-1e15-44cf-87f1-6037f6631592-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.233034 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dd2bq\" (UniqueName: \"kubernetes.io/projected/226b6f72-5676-4dc8-9536-d99da01652c8-kube-api-access-dd2bq\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.233054 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fvpq\" (UniqueName: \"kubernetes.io/projected/2235e98c-f4f9-40ae-8f47-788c336e5893-kube-api-access-9fvpq\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.233065 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2235e98c-f4f9-40ae-8f47-788c336e5893-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.233078 4857 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01903422-1e15-44cf-87f1-6037f6631592-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.233087 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2235e98c-f4f9-40ae-8f47-788c336e5893-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.233095 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ng9h\" (UniqueName: \"kubernetes.io/projected/01903422-1e15-44cf-87f1-6037f6631592-kube-api-access-5ng9h\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.243132 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4s5mw"] Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.253686 4857 scope.go:117] "RemoveContainer" containerID="416d990ea4600370647de5c70a400a0db0f4bcc98d93519e4132ed021ef0f2e4" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.254879 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4s5mw"] Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.262143 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mbzs2"] Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.265238 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mbzs2"] Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.267663 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/226b6f72-5676-4dc8-9536-d99da01652c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "226b6f72-5676-4dc8-9536-d99da01652c8" (UID: "226b6f72-5676-4dc8-9536-d99da01652c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.268275 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-thg27"] Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.270127 4857 scope.go:117] "RemoveContainer" containerID="511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.270477 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5\": container with ID starting with 511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5 not found: ID does not exist" containerID="511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.270518 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5"} err="failed to get container status \"511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5\": rpc error: code = NotFound desc = could not find container \"511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5\": container with ID starting with 511226207debdb8bfe094fdfbc41cc1fed5aee283a8e68cda66625a999de7dd5 not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.270545 4857 scope.go:117] "RemoveContainer" containerID="4330ab6c16a16ed62e8ea9df65b84cbf1e29ab7f3560c278f9884f8ab04914a2" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.270710 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4330ab6c16a16ed62e8ea9df65b84cbf1e29ab7f3560c278f9884f8ab04914a2\": container with ID starting with 4330ab6c16a16ed62e8ea9df65b84cbf1e29ab7f3560c278f9884f8ab04914a2 not found: ID does not exist" containerID="4330ab6c16a16ed62e8ea9df65b84cbf1e29ab7f3560c278f9884f8ab04914a2" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.270737 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4330ab6c16a16ed62e8ea9df65b84cbf1e29ab7f3560c278f9884f8ab04914a2"} err="failed to get container status \"4330ab6c16a16ed62e8ea9df65b84cbf1e29ab7f3560c278f9884f8ab04914a2\": rpc error: code = NotFound desc = could not find container \"4330ab6c16a16ed62e8ea9df65b84cbf1e29ab7f3560c278f9884f8ab04914a2\": container with ID starting with 4330ab6c16a16ed62e8ea9df65b84cbf1e29ab7f3560c278f9884f8ab04914a2 not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.270751 4857 scope.go:117] "RemoveContainer" containerID="416d990ea4600370647de5c70a400a0db0f4bcc98d93519e4132ed021ef0f2e4" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.270899 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"416d990ea4600370647de5c70a400a0db0f4bcc98d93519e4132ed021ef0f2e4\": container with ID starting with 416d990ea4600370647de5c70a400a0db0f4bcc98d93519e4132ed021ef0f2e4 not found: ID does not exist" containerID="416d990ea4600370647de5c70a400a0db0f4bcc98d93519e4132ed021ef0f2e4" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.270920 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"416d990ea4600370647de5c70a400a0db0f4bcc98d93519e4132ed021ef0f2e4"} err="failed to get container status \"416d990ea4600370647de5c70a400a0db0f4bcc98d93519e4132ed021ef0f2e4\": rpc error: code = NotFound desc = could not find container \"416d990ea4600370647de5c70a400a0db0f4bcc98d93519e4132ed021ef0f2e4\": container with ID starting with 416d990ea4600370647de5c70a400a0db0f4bcc98d93519e4132ed021ef0f2e4 not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.270932 4857 scope.go:117] "RemoveContainer" containerID="76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.271957 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-thg27"] Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.285387 4857 scope.go:117] "RemoveContainer" containerID="acdb4cdaf26d61dc3fd6b5cf99496a03cce2574a9542a00a1f4caf3858e2702b" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.299863 4857 scope.go:117] "RemoveContainer" containerID="76a33f14ae5a7f7cca7acbe239f2b8e6300df462ba9bc33a262fcc1c2912ad47" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.309239 4857 scope.go:117] "RemoveContainer" containerID="76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.309507 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56\": container with ID starting with 76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56 not found: ID does not exist" containerID="76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.309548 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56"} err="failed to get container status \"76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56\": rpc error: code = NotFound desc = could not find container \"76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56\": container with ID starting with 76f714bebc38f795d47338a9604e9b14bc000e9bd6dd6954db56d25d3059bf56 not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.309575 4857 scope.go:117] "RemoveContainer" containerID="acdb4cdaf26d61dc3fd6b5cf99496a03cce2574a9542a00a1f4caf3858e2702b" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.309824 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acdb4cdaf26d61dc3fd6b5cf99496a03cce2574a9542a00a1f4caf3858e2702b\": container with ID starting with acdb4cdaf26d61dc3fd6b5cf99496a03cce2574a9542a00a1f4caf3858e2702b not found: ID does not exist" containerID="acdb4cdaf26d61dc3fd6b5cf99496a03cce2574a9542a00a1f4caf3858e2702b" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.309852 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acdb4cdaf26d61dc3fd6b5cf99496a03cce2574a9542a00a1f4caf3858e2702b"} err="failed to get container status \"acdb4cdaf26d61dc3fd6b5cf99496a03cce2574a9542a00a1f4caf3858e2702b\": rpc error: code = NotFound desc = could not find container \"acdb4cdaf26d61dc3fd6b5cf99496a03cce2574a9542a00a1f4caf3858e2702b\": container with ID starting with acdb4cdaf26d61dc3fd6b5cf99496a03cce2574a9542a00a1f4caf3858e2702b not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.309876 4857 scope.go:117] "RemoveContainer" containerID="76a33f14ae5a7f7cca7acbe239f2b8e6300df462ba9bc33a262fcc1c2912ad47" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.310125 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76a33f14ae5a7f7cca7acbe239f2b8e6300df462ba9bc33a262fcc1c2912ad47\": container with ID starting with 76a33f14ae5a7f7cca7acbe239f2b8e6300df462ba9bc33a262fcc1c2912ad47 not found: ID does not exist" containerID="76a33f14ae5a7f7cca7acbe239f2b8e6300df462ba9bc33a262fcc1c2912ad47" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.310149 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76a33f14ae5a7f7cca7acbe239f2b8e6300df462ba9bc33a262fcc1c2912ad47"} err="failed to get container status \"76a33f14ae5a7f7cca7acbe239f2b8e6300df462ba9bc33a262fcc1c2912ad47\": rpc error: code = NotFound desc = could not find container \"76a33f14ae5a7f7cca7acbe239f2b8e6300df462ba9bc33a262fcc1c2912ad47\": container with ID starting with 76a33f14ae5a7f7cca7acbe239f2b8e6300df462ba9bc33a262fcc1c2912ad47 not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.310162 4857 scope.go:117] "RemoveContainer" containerID="d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.320568 4857 scope.go:117] "RemoveContainer" containerID="9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.333552 4857 scope.go:117] "RemoveContainer" containerID="d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.334365 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef\": container with ID starting with d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef not found: ID does not exist" containerID="d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.334409 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef"} err="failed to get container status \"d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef\": rpc error: code = NotFound desc = could not find container \"d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef\": container with ID starting with d5606228c62de65eb5096d12fcdbcadb143733b9b6cb4bcd8b90fcf08e8e9eef not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.334438 4857 scope.go:117] "RemoveContainer" containerID="9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.334687 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/226b6f72-5676-4dc8-9536-d99da01652c8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.334860 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a\": container with ID starting with 9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a not found: ID does not exist" containerID="9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.334903 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a"} err="failed to get container status \"9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a\": rpc error: code = NotFound desc = could not find container \"9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a\": container with ID starting with 9e6e71147bc17f6421e993921c49e21f82a5ef3b8c2c7f831c2c01e0fc7df67a not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.334932 4857 scope.go:117] "RemoveContainer" containerID="6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.346002 4857 scope.go:117] "RemoveContainer" containerID="bfbfc9af4f9f8fa180d661e2bab33cb070216b92e544c8487a17baee12962879" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.363537 4857 scope.go:117] "RemoveContainer" containerID="58e1feea57545e597fcbc11f62041bd0f69a2f9188d4f5d9404716bde32e7c17" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.377083 4857 scope.go:117] "RemoveContainer" containerID="6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.377416 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f\": container with ID starting with 6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f not found: ID does not exist" containerID="6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.377446 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f"} err="failed to get container status \"6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f\": rpc error: code = NotFound desc = could not find container \"6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f\": container with ID starting with 6f3c6b5c134cb39ab73fe37e228bebb4bf8d17c5a001119dc77419f6da53612f not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.377482 4857 scope.go:117] "RemoveContainer" containerID="bfbfc9af4f9f8fa180d661e2bab33cb070216b92e544c8487a17baee12962879" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.378070 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfbfc9af4f9f8fa180d661e2bab33cb070216b92e544c8487a17baee12962879\": container with ID starting with bfbfc9af4f9f8fa180d661e2bab33cb070216b92e544c8487a17baee12962879 not found: ID does not exist" containerID="bfbfc9af4f9f8fa180d661e2bab33cb070216b92e544c8487a17baee12962879" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.378112 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfbfc9af4f9f8fa180d661e2bab33cb070216b92e544c8487a17baee12962879"} err="failed to get container status \"bfbfc9af4f9f8fa180d661e2bab33cb070216b92e544c8487a17baee12962879\": rpc error: code = NotFound desc = could not find container \"bfbfc9af4f9f8fa180d661e2bab33cb070216b92e544c8487a17baee12962879\": container with ID starting with bfbfc9af4f9f8fa180d661e2bab33cb070216b92e544c8487a17baee12962879 not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.378143 4857 scope.go:117] "RemoveContainer" containerID="58e1feea57545e597fcbc11f62041bd0f69a2f9188d4f5d9404716bde32e7c17" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.378398 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58e1feea57545e597fcbc11f62041bd0f69a2f9188d4f5d9404716bde32e7c17\": container with ID starting with 58e1feea57545e597fcbc11f62041bd0f69a2f9188d4f5d9404716bde32e7c17 not found: ID does not exist" containerID="58e1feea57545e597fcbc11f62041bd0f69a2f9188d4f5d9404716bde32e7c17" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.378416 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58e1feea57545e597fcbc11f62041bd0f69a2f9188d4f5d9404716bde32e7c17"} err="failed to get container status \"58e1feea57545e597fcbc11f62041bd0f69a2f9188d4f5d9404716bde32e7c17\": rpc error: code = NotFound desc = could not find container \"58e1feea57545e597fcbc11f62041bd0f69a2f9188d4f5d9404716bde32e7c17\": container with ID starting with 58e1feea57545e597fcbc11f62041bd0f69a2f9188d4f5d9404716bde32e7c17 not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.378430 4857 scope.go:117] "RemoveContainer" containerID="f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.388497 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d5gjm"] Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.392751 4857 scope.go:117] "RemoveContainer" containerID="ba8eeae978be367d4518757a0fee5ecc88932ebd5f3bcc7f17aaafae99638ef9" Dec 01 21:41:19 crc kubenswrapper[4857]: W1201 21:41:19.396901 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fe4a11e_f252_4f61_a8f4_d513b2b40643.slice/crio-a7bbfe0029434257e14d428931965d6f88cc17617b2a2e9f0a57a2b48c2b7961 WatchSource:0}: Error finding container a7bbfe0029434257e14d428931965d6f88cc17617b2a2e9f0a57a2b48c2b7961: Status 404 returned error can't find the container with id a7bbfe0029434257e14d428931965d6f88cc17617b2a2e9f0a57a2b48c2b7961 Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.409173 4857 scope.go:117] "RemoveContainer" containerID="2f9577782abbc8097923870aa3f701119563192fc8c56de31690da853351ee3c" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.423908 4857 scope.go:117] "RemoveContainer" containerID="f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.424236 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf\": container with ID starting with f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf not found: ID does not exist" containerID="f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.424281 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf"} err="failed to get container status \"f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf\": rpc error: code = NotFound desc = could not find container \"f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf\": container with ID starting with f7709ec44c8dc3ded823c0629d6003dc05c9d2572a6c0a44f0308629a39632cf not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.424302 4857 scope.go:117] "RemoveContainer" containerID="ba8eeae978be367d4518757a0fee5ecc88932ebd5f3bcc7f17aaafae99638ef9" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.424563 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba8eeae978be367d4518757a0fee5ecc88932ebd5f3bcc7f17aaafae99638ef9\": container with ID starting with ba8eeae978be367d4518757a0fee5ecc88932ebd5f3bcc7f17aaafae99638ef9 not found: ID does not exist" containerID="ba8eeae978be367d4518757a0fee5ecc88932ebd5f3bcc7f17aaafae99638ef9" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.424624 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba8eeae978be367d4518757a0fee5ecc88932ebd5f3bcc7f17aaafae99638ef9"} err="failed to get container status \"ba8eeae978be367d4518757a0fee5ecc88932ebd5f3bcc7f17aaafae99638ef9\": rpc error: code = NotFound desc = could not find container \"ba8eeae978be367d4518757a0fee5ecc88932ebd5f3bcc7f17aaafae99638ef9\": container with ID starting with ba8eeae978be367d4518757a0fee5ecc88932ebd5f3bcc7f17aaafae99638ef9 not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.424660 4857 scope.go:117] "RemoveContainer" containerID="2f9577782abbc8097923870aa3f701119563192fc8c56de31690da853351ee3c" Dec 01 21:41:19 crc kubenswrapper[4857]: E1201 21:41:19.431478 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f9577782abbc8097923870aa3f701119563192fc8c56de31690da853351ee3c\": container with ID starting with 2f9577782abbc8097923870aa3f701119563192fc8c56de31690da853351ee3c not found: ID does not exist" containerID="2f9577782abbc8097923870aa3f701119563192fc8c56de31690da853351ee3c" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.431508 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f9577782abbc8097923870aa3f701119563192fc8c56de31690da853351ee3c"} err="failed to get container status \"2f9577782abbc8097923870aa3f701119563192fc8c56de31690da853351ee3c\": rpc error: code = NotFound desc = could not find container \"2f9577782abbc8097923870aa3f701119563192fc8c56de31690da853351ee3c\": container with ID starting with 2f9577782abbc8097923870aa3f701119563192fc8c56de31690da853351ee3c not found: ID does not exist" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.525681 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zrwmc"] Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.528924 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zrwmc"] Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.842668 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01903422-1e15-44cf-87f1-6037f6631592" path="/var/lib/kubelet/pods/01903422-1e15-44cf-87f1-6037f6631592/volumes" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.844001 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2235e98c-f4f9-40ae-8f47-788c336e5893" path="/var/lib/kubelet/pods/2235e98c-f4f9-40ae-8f47-788c336e5893/volumes" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.845286 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="226b6f72-5676-4dc8-9536-d99da01652c8" path="/var/lib/kubelet/pods/226b6f72-5676-4dc8-9536-d99da01652c8/volumes" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.847372 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" path="/var/lib/kubelet/pods/2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a/volumes" Dec 01 21:41:19 crc kubenswrapper[4857]: I1201 21:41:19.848632 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8049f37f-5448-4796-b384-691649bd559c" path="/var/lib/kubelet/pods/8049f37f-5448-4796-b384-691649bd559c/volumes" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.199242 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" event={"ID":"9fe4a11e-f252-4f61-a8f4-d513b2b40643","Type":"ContainerStarted","Data":"39fddb41cce5975e12cf8d405f71cdf023105903e12feb52c22e3c9e4b4a9bd5"} Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.199662 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" event={"ID":"9fe4a11e-f252-4f61-a8f4-d513b2b40643","Type":"ContainerStarted","Data":"a7bbfe0029434257e14d428931965d6f88cc17617b2a2e9f0a57a2b48c2b7961"} Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.199945 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.206341 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.216768 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-d5gjm" podStartSLOduration=2.216749093 podStartE2EDuration="2.216749093s" podCreationTimestamp="2025-12-01 21:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:41:20.213269311 +0000 UTC m=+398.703331658" watchObservedRunningTime="2025-12-01 21:41:20.216749093 +0000 UTC m=+398.706811410" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.596066 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2x8zg"] Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.597430 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2235e98c-f4f9-40ae-8f47-788c336e5893" containerName="extract-content" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.597488 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="2235e98c-f4f9-40ae-8f47-788c336e5893" containerName="extract-content" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.597528 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01903422-1e15-44cf-87f1-6037f6631592" containerName="marketplace-operator" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.597546 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="01903422-1e15-44cf-87f1-6037f6631592" containerName="marketplace-operator" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.597595 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2235e98c-f4f9-40ae-8f47-788c336e5893" containerName="registry-server" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.597629 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="2235e98c-f4f9-40ae-8f47-788c336e5893" containerName="registry-server" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.597658 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8049f37f-5448-4796-b384-691649bd559c" containerName="extract-utilities" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.597672 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8049f37f-5448-4796-b384-691649bd559c" containerName="extract-utilities" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.597719 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" containerName="extract-content" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.597732 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" containerName="extract-content" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.597769 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" containerName="registry-server" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.597788 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" containerName="registry-server" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.597818 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="226b6f72-5676-4dc8-9536-d99da01652c8" containerName="extract-content" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.597834 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="226b6f72-5676-4dc8-9536-d99da01652c8" containerName="extract-content" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.597855 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8049f37f-5448-4796-b384-691649bd559c" containerName="registry-server" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.597875 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8049f37f-5448-4796-b384-691649bd559c" containerName="registry-server" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.597895 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8049f37f-5448-4796-b384-691649bd559c" containerName="extract-content" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.597908 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8049f37f-5448-4796-b384-691649bd559c" containerName="extract-content" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.597935 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="226b6f72-5676-4dc8-9536-d99da01652c8" containerName="extract-utilities" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.597948 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="226b6f72-5676-4dc8-9536-d99da01652c8" containerName="extract-utilities" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.597982 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2235e98c-f4f9-40ae-8f47-788c336e5893" containerName="extract-utilities" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.597995 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="2235e98c-f4f9-40ae-8f47-788c336e5893" containerName="extract-utilities" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.598022 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="226b6f72-5676-4dc8-9536-d99da01652c8" containerName="registry-server" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.598036 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="226b6f72-5676-4dc8-9536-d99da01652c8" containerName="registry-server" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.598087 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" containerName="extract-utilities" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.598125 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" containerName="extract-utilities" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.598661 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b2aa615-5e1e-47b6-8e23-3cd3cf0f328a" containerName="registry-server" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.598734 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="8049f37f-5448-4796-b384-691649bd559c" containerName="registry-server" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.598750 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="2235e98c-f4f9-40ae-8f47-788c336e5893" containerName="registry-server" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.598765 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="226b6f72-5676-4dc8-9536-d99da01652c8" containerName="registry-server" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.598796 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="01903422-1e15-44cf-87f1-6037f6631592" containerName="marketplace-operator" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.598817 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="01903422-1e15-44cf-87f1-6037f6631592" containerName="marketplace-operator" Dec 01 21:41:20 crc kubenswrapper[4857]: E1201 21:41:20.599209 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01903422-1e15-44cf-87f1-6037f6631592" containerName="marketplace-operator" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.599235 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="01903422-1e15-44cf-87f1-6037f6631592" containerName="marketplace-operator" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.606977 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.608268 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2x8zg"] Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.613612 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.757519 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c20d9041-cb65-4743-ac7b-3f0f4675b71e-catalog-content\") pod \"redhat-marketplace-2x8zg\" (UID: \"c20d9041-cb65-4743-ac7b-3f0f4675b71e\") " pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.757579 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8w6n\" (UniqueName: \"kubernetes.io/projected/c20d9041-cb65-4743-ac7b-3f0f4675b71e-kube-api-access-x8w6n\") pod \"redhat-marketplace-2x8zg\" (UID: \"c20d9041-cb65-4743-ac7b-3f0f4675b71e\") " pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.757608 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c20d9041-cb65-4743-ac7b-3f0f4675b71e-utilities\") pod \"redhat-marketplace-2x8zg\" (UID: \"c20d9041-cb65-4743-ac7b-3f0f4675b71e\") " pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.785088 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zfmgr"] Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.786845 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.791940 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.803730 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zfmgr"] Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.859256 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b7910c3-1306-4e26-a5d6-95a78e751adc-catalog-content\") pod \"redhat-operators-zfmgr\" (UID: \"0b7910c3-1306-4e26-a5d6-95a78e751adc\") " pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.859362 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6knfg\" (UniqueName: \"kubernetes.io/projected/0b7910c3-1306-4e26-a5d6-95a78e751adc-kube-api-access-6knfg\") pod \"redhat-operators-zfmgr\" (UID: \"0b7910c3-1306-4e26-a5d6-95a78e751adc\") " pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.859487 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c20d9041-cb65-4743-ac7b-3f0f4675b71e-catalog-content\") pod \"redhat-marketplace-2x8zg\" (UID: \"c20d9041-cb65-4743-ac7b-3f0f4675b71e\") " pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.859541 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8w6n\" (UniqueName: \"kubernetes.io/projected/c20d9041-cb65-4743-ac7b-3f0f4675b71e-kube-api-access-x8w6n\") pod \"redhat-marketplace-2x8zg\" (UID: \"c20d9041-cb65-4743-ac7b-3f0f4675b71e\") " pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.859597 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b7910c3-1306-4e26-a5d6-95a78e751adc-utilities\") pod \"redhat-operators-zfmgr\" (UID: \"0b7910c3-1306-4e26-a5d6-95a78e751adc\") " pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.859756 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c20d9041-cb65-4743-ac7b-3f0f4675b71e-utilities\") pod \"redhat-marketplace-2x8zg\" (UID: \"c20d9041-cb65-4743-ac7b-3f0f4675b71e\") " pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.860683 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c20d9041-cb65-4743-ac7b-3f0f4675b71e-utilities\") pod \"redhat-marketplace-2x8zg\" (UID: \"c20d9041-cb65-4743-ac7b-3f0f4675b71e\") " pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.861929 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c20d9041-cb65-4743-ac7b-3f0f4675b71e-catalog-content\") pod \"redhat-marketplace-2x8zg\" (UID: \"c20d9041-cb65-4743-ac7b-3f0f4675b71e\") " pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.889525 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8w6n\" (UniqueName: \"kubernetes.io/projected/c20d9041-cb65-4743-ac7b-3f0f4675b71e-kube-api-access-x8w6n\") pod \"redhat-marketplace-2x8zg\" (UID: \"c20d9041-cb65-4743-ac7b-3f0f4675b71e\") " pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.932539 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.961787 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b7910c3-1306-4e26-a5d6-95a78e751adc-utilities\") pod \"redhat-operators-zfmgr\" (UID: \"0b7910c3-1306-4e26-a5d6-95a78e751adc\") " pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.961936 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b7910c3-1306-4e26-a5d6-95a78e751adc-catalog-content\") pod \"redhat-operators-zfmgr\" (UID: \"0b7910c3-1306-4e26-a5d6-95a78e751adc\") " pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.961994 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6knfg\" (UniqueName: \"kubernetes.io/projected/0b7910c3-1306-4e26-a5d6-95a78e751adc-kube-api-access-6knfg\") pod \"redhat-operators-zfmgr\" (UID: \"0b7910c3-1306-4e26-a5d6-95a78e751adc\") " pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.962886 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b7910c3-1306-4e26-a5d6-95a78e751adc-catalog-content\") pod \"redhat-operators-zfmgr\" (UID: \"0b7910c3-1306-4e26-a5d6-95a78e751adc\") " pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.962966 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b7910c3-1306-4e26-a5d6-95a78e751adc-utilities\") pod \"redhat-operators-zfmgr\" (UID: \"0b7910c3-1306-4e26-a5d6-95a78e751adc\") " pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:20 crc kubenswrapper[4857]: I1201 21:41:20.988986 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6knfg\" (UniqueName: \"kubernetes.io/projected/0b7910c3-1306-4e26-a5d6-95a78e751adc-kube-api-access-6knfg\") pod \"redhat-operators-zfmgr\" (UID: \"0b7910c3-1306-4e26-a5d6-95a78e751adc\") " pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:21 crc kubenswrapper[4857]: I1201 21:41:21.105424 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:21 crc kubenswrapper[4857]: I1201 21:41:21.365284 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zfmgr"] Dec 01 21:41:21 crc kubenswrapper[4857]: W1201 21:41:21.370287 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b7910c3_1306_4e26_a5d6_95a78e751adc.slice/crio-61af1c1e8570c92a160d4f19a699461b3e3bcf47c4bd29383a60c6c166762218 WatchSource:0}: Error finding container 61af1c1e8570c92a160d4f19a699461b3e3bcf47c4bd29383a60c6c166762218: Status 404 returned error can't find the container with id 61af1c1e8570c92a160d4f19a699461b3e3bcf47c4bd29383a60c6c166762218 Dec 01 21:41:21 crc kubenswrapper[4857]: I1201 21:41:21.384769 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2x8zg"] Dec 01 21:41:21 crc kubenswrapper[4857]: W1201 21:41:21.392492 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc20d9041_cb65_4743_ac7b_3f0f4675b71e.slice/crio-79f5c89f69200544d10e1158661c88709cf0ac819b26d299307fe40e31c47ad1 WatchSource:0}: Error finding container 79f5c89f69200544d10e1158661c88709cf0ac819b26d299307fe40e31c47ad1: Status 404 returned error can't find the container with id 79f5c89f69200544d10e1158661c88709cf0ac819b26d299307fe40e31c47ad1 Dec 01 21:41:22 crc kubenswrapper[4857]: I1201 21:41:22.223452 4857 generic.go:334] "Generic (PLEG): container finished" podID="0b7910c3-1306-4e26-a5d6-95a78e751adc" containerID="c745f30b6632b71bfc94a63db87a0024c00b6e72eae7239772288e42ac435381" exitCode=0 Dec 01 21:41:22 crc kubenswrapper[4857]: I1201 21:41:22.223552 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfmgr" event={"ID":"0b7910c3-1306-4e26-a5d6-95a78e751adc","Type":"ContainerDied","Data":"c745f30b6632b71bfc94a63db87a0024c00b6e72eae7239772288e42ac435381"} Dec 01 21:41:22 crc kubenswrapper[4857]: I1201 21:41:22.223993 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfmgr" event={"ID":"0b7910c3-1306-4e26-a5d6-95a78e751adc","Type":"ContainerStarted","Data":"61af1c1e8570c92a160d4f19a699461b3e3bcf47c4bd29383a60c6c166762218"} Dec 01 21:41:22 crc kubenswrapper[4857]: I1201 21:41:22.226115 4857 generic.go:334] "Generic (PLEG): container finished" podID="c20d9041-cb65-4743-ac7b-3f0f4675b71e" containerID="ec1bd2389616ae02a89f9071b3e839101206bf6ce998cfd6cc5180edb34c1cea" exitCode=0 Dec 01 21:41:22 crc kubenswrapper[4857]: I1201 21:41:22.226311 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2x8zg" event={"ID":"c20d9041-cb65-4743-ac7b-3f0f4675b71e","Type":"ContainerDied","Data":"ec1bd2389616ae02a89f9071b3e839101206bf6ce998cfd6cc5180edb34c1cea"} Dec 01 21:41:22 crc kubenswrapper[4857]: I1201 21:41:22.226376 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2x8zg" event={"ID":"c20d9041-cb65-4743-ac7b-3f0f4675b71e","Type":"ContainerStarted","Data":"79f5c89f69200544d10e1158661c88709cf0ac819b26d299307fe40e31c47ad1"} Dec 01 21:41:22 crc kubenswrapper[4857]: I1201 21:41:22.983570 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d8qfd"] Dec 01 21:41:22 crc kubenswrapper[4857]: I1201 21:41:22.986463 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:22 crc kubenswrapper[4857]: I1201 21:41:22.991592 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 21:41:22 crc kubenswrapper[4857]: I1201 21:41:22.995870 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d8qfd"] Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.091945 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3e413ef-6fbc-4e78-8650-357ea026b79c-utilities\") pod \"certified-operators-d8qfd\" (UID: \"c3e413ef-6fbc-4e78-8650-357ea026b79c\") " pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.092356 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqrq4\" (UniqueName: \"kubernetes.io/projected/c3e413ef-6fbc-4e78-8650-357ea026b79c-kube-api-access-zqrq4\") pod \"certified-operators-d8qfd\" (UID: \"c3e413ef-6fbc-4e78-8650-357ea026b79c\") " pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.092410 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3e413ef-6fbc-4e78-8650-357ea026b79c-catalog-content\") pod \"certified-operators-d8qfd\" (UID: \"c3e413ef-6fbc-4e78-8650-357ea026b79c\") " pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.173908 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mw8ff"] Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.175558 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.182270 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.189358 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mw8ff"] Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.198531 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqrq4\" (UniqueName: \"kubernetes.io/projected/c3e413ef-6fbc-4e78-8650-357ea026b79c-kube-api-access-zqrq4\") pod \"certified-operators-d8qfd\" (UID: \"c3e413ef-6fbc-4e78-8650-357ea026b79c\") " pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.198590 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3e413ef-6fbc-4e78-8650-357ea026b79c-catalog-content\") pod \"certified-operators-d8qfd\" (UID: \"c3e413ef-6fbc-4e78-8650-357ea026b79c\") " pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.198667 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3e413ef-6fbc-4e78-8650-357ea026b79c-utilities\") pod \"certified-operators-d8qfd\" (UID: \"c3e413ef-6fbc-4e78-8650-357ea026b79c\") " pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.201709 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3e413ef-6fbc-4e78-8650-357ea026b79c-catalog-content\") pod \"certified-operators-d8qfd\" (UID: \"c3e413ef-6fbc-4e78-8650-357ea026b79c\") " pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.201898 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3e413ef-6fbc-4e78-8650-357ea026b79c-utilities\") pod \"certified-operators-d8qfd\" (UID: \"c3e413ef-6fbc-4e78-8650-357ea026b79c\") " pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.228172 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqrq4\" (UniqueName: \"kubernetes.io/projected/c3e413ef-6fbc-4e78-8650-357ea026b79c-kube-api-access-zqrq4\") pod \"certified-operators-d8qfd\" (UID: \"c3e413ef-6fbc-4e78-8650-357ea026b79c\") " pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.234847 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfmgr" event={"ID":"0b7910c3-1306-4e26-a5d6-95a78e751adc","Type":"ContainerStarted","Data":"dddb5d0af4764b03c49dc2fddb032ddc6a52f7675a3244360335445dbd373207"} Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.237961 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2x8zg" event={"ID":"c20d9041-cb65-4743-ac7b-3f0f4675b71e","Type":"ContainerStarted","Data":"76138374981e1ac1da4dfbd79aa002209bf5ce69b0be9ef2f8e88b229a6eab5e"} Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.299641 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvckk\" (UniqueName: \"kubernetes.io/projected/00a902c7-b8d4-4f4d-bc21-875579a6f0d0-kube-api-access-gvckk\") pod \"community-operators-mw8ff\" (UID: \"00a902c7-b8d4-4f4d-bc21-875579a6f0d0\") " pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.299694 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a902c7-b8d4-4f4d-bc21-875579a6f0d0-catalog-content\") pod \"community-operators-mw8ff\" (UID: \"00a902c7-b8d4-4f4d-bc21-875579a6f0d0\") " pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.299769 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a902c7-b8d4-4f4d-bc21-875579a6f0d0-utilities\") pod \"community-operators-mw8ff\" (UID: \"00a902c7-b8d4-4f4d-bc21-875579a6f0d0\") " pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.361662 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.401294 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a902c7-b8d4-4f4d-bc21-875579a6f0d0-utilities\") pod \"community-operators-mw8ff\" (UID: \"00a902c7-b8d4-4f4d-bc21-875579a6f0d0\") " pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.401399 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvckk\" (UniqueName: \"kubernetes.io/projected/00a902c7-b8d4-4f4d-bc21-875579a6f0d0-kube-api-access-gvckk\") pod \"community-operators-mw8ff\" (UID: \"00a902c7-b8d4-4f4d-bc21-875579a6f0d0\") " pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.401437 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a902c7-b8d4-4f4d-bc21-875579a6f0d0-catalog-content\") pod \"community-operators-mw8ff\" (UID: \"00a902c7-b8d4-4f4d-bc21-875579a6f0d0\") " pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.402282 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a902c7-b8d4-4f4d-bc21-875579a6f0d0-utilities\") pod \"community-operators-mw8ff\" (UID: \"00a902c7-b8d4-4f4d-bc21-875579a6f0d0\") " pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.402439 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a902c7-b8d4-4f4d-bc21-875579a6f0d0-catalog-content\") pod \"community-operators-mw8ff\" (UID: \"00a902c7-b8d4-4f4d-bc21-875579a6f0d0\") " pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.427210 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvckk\" (UniqueName: \"kubernetes.io/projected/00a902c7-b8d4-4f4d-bc21-875579a6f0d0-kube-api-access-gvckk\") pod \"community-operators-mw8ff\" (UID: \"00a902c7-b8d4-4f4d-bc21-875579a6f0d0\") " pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.491401 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.594504 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d8qfd"] Dec 01 21:41:23 crc kubenswrapper[4857]: W1201 21:41:23.598775 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3e413ef_6fbc_4e78_8650_357ea026b79c.slice/crio-d251a22e99d9d3f9a38087cc691fb8175377fe9eff3cccfb31f374cdeb8dd568 WatchSource:0}: Error finding container d251a22e99d9d3f9a38087cc691fb8175377fe9eff3cccfb31f374cdeb8dd568: Status 404 returned error can't find the container with id d251a22e99d9d3f9a38087cc691fb8175377fe9eff3cccfb31f374cdeb8dd568 Dec 01 21:41:23 crc kubenswrapper[4857]: I1201 21:41:23.711674 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mw8ff"] Dec 01 21:41:23 crc kubenswrapper[4857]: W1201 21:41:23.803179 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00a902c7_b8d4_4f4d_bc21_875579a6f0d0.slice/crio-0e3cc7eb9fd009abfeb89254420a6eba7fc2dd50b67993e157fe874382f6cef9 WatchSource:0}: Error finding container 0e3cc7eb9fd009abfeb89254420a6eba7fc2dd50b67993e157fe874382f6cef9: Status 404 returned error can't find the container with id 0e3cc7eb9fd009abfeb89254420a6eba7fc2dd50b67993e157fe874382f6cef9 Dec 01 21:41:24 crc kubenswrapper[4857]: I1201 21:41:24.246277 4857 generic.go:334] "Generic (PLEG): container finished" podID="00a902c7-b8d4-4f4d-bc21-875579a6f0d0" containerID="97b35ef065c0dcc6d8d9bd6fc8b385db6ca23473f1c81d1996c1029481d5793c" exitCode=0 Dec 01 21:41:24 crc kubenswrapper[4857]: I1201 21:41:24.246384 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mw8ff" event={"ID":"00a902c7-b8d4-4f4d-bc21-875579a6f0d0","Type":"ContainerDied","Data":"97b35ef065c0dcc6d8d9bd6fc8b385db6ca23473f1c81d1996c1029481d5793c"} Dec 01 21:41:24 crc kubenswrapper[4857]: I1201 21:41:24.246424 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mw8ff" event={"ID":"00a902c7-b8d4-4f4d-bc21-875579a6f0d0","Type":"ContainerStarted","Data":"0e3cc7eb9fd009abfeb89254420a6eba7fc2dd50b67993e157fe874382f6cef9"} Dec 01 21:41:24 crc kubenswrapper[4857]: I1201 21:41:24.251229 4857 generic.go:334] "Generic (PLEG): container finished" podID="c20d9041-cb65-4743-ac7b-3f0f4675b71e" containerID="76138374981e1ac1da4dfbd79aa002209bf5ce69b0be9ef2f8e88b229a6eab5e" exitCode=0 Dec 01 21:41:24 crc kubenswrapper[4857]: I1201 21:41:24.251319 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2x8zg" event={"ID":"c20d9041-cb65-4743-ac7b-3f0f4675b71e","Type":"ContainerDied","Data":"76138374981e1ac1da4dfbd79aa002209bf5ce69b0be9ef2f8e88b229a6eab5e"} Dec 01 21:41:24 crc kubenswrapper[4857]: I1201 21:41:24.254976 4857 generic.go:334] "Generic (PLEG): container finished" podID="0b7910c3-1306-4e26-a5d6-95a78e751adc" containerID="dddb5d0af4764b03c49dc2fddb032ddc6a52f7675a3244360335445dbd373207" exitCode=0 Dec 01 21:41:24 crc kubenswrapper[4857]: I1201 21:41:24.255033 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfmgr" event={"ID":"0b7910c3-1306-4e26-a5d6-95a78e751adc","Type":"ContainerDied","Data":"dddb5d0af4764b03c49dc2fddb032ddc6a52f7675a3244360335445dbd373207"} Dec 01 21:41:24 crc kubenswrapper[4857]: I1201 21:41:24.261991 4857 generic.go:334] "Generic (PLEG): container finished" podID="c3e413ef-6fbc-4e78-8650-357ea026b79c" containerID="0e244840b02dc2001eb5a3e1418441db8cd6a1aa705ef0dd6f0d8bbd3da43a9d" exitCode=0 Dec 01 21:41:24 crc kubenswrapper[4857]: I1201 21:41:24.262214 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8qfd" event={"ID":"c3e413ef-6fbc-4e78-8650-357ea026b79c","Type":"ContainerDied","Data":"0e244840b02dc2001eb5a3e1418441db8cd6a1aa705ef0dd6f0d8bbd3da43a9d"} Dec 01 21:41:24 crc kubenswrapper[4857]: I1201 21:41:24.262386 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8qfd" event={"ID":"c3e413ef-6fbc-4e78-8650-357ea026b79c","Type":"ContainerStarted","Data":"d251a22e99d9d3f9a38087cc691fb8175377fe9eff3cccfb31f374cdeb8dd568"} Dec 01 21:41:25 crc kubenswrapper[4857]: I1201 21:41:25.271460 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2x8zg" event={"ID":"c20d9041-cb65-4743-ac7b-3f0f4675b71e","Type":"ContainerStarted","Data":"68f3d5f0b62917619079ea49e4851d37716a0b2839c2a8c1a1c452107546214b"} Dec 01 21:41:25 crc kubenswrapper[4857]: I1201 21:41:25.282082 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfmgr" event={"ID":"0b7910c3-1306-4e26-a5d6-95a78e751adc","Type":"ContainerStarted","Data":"c89cb496d28eb32f135e2056c9bfa78ed4bda8c95feb0f1f4453a9795032bbad"} Dec 01 21:41:25 crc kubenswrapper[4857]: I1201 21:41:25.284978 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8qfd" event={"ID":"c3e413ef-6fbc-4e78-8650-357ea026b79c","Type":"ContainerStarted","Data":"136e8a2a11e28a131644fa0b7b636f7eabd24e06d37059d5c345547105f95362"} Dec 01 21:41:25 crc kubenswrapper[4857]: I1201 21:41:25.300941 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2x8zg" podStartSLOduration=2.705683981 podStartE2EDuration="5.300900737s" podCreationTimestamp="2025-12-01 21:41:20 +0000 UTC" firstStartedPulling="2025-12-01 21:41:22.227887306 +0000 UTC m=+400.717949663" lastFinishedPulling="2025-12-01 21:41:24.823104102 +0000 UTC m=+403.313166419" observedRunningTime="2025-12-01 21:41:25.300451063 +0000 UTC m=+403.790513380" watchObservedRunningTime="2025-12-01 21:41:25.300900737 +0000 UTC m=+403.790963054" Dec 01 21:41:25 crc kubenswrapper[4857]: I1201 21:41:25.341334 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zfmgr" podStartSLOduration=2.827692489 podStartE2EDuration="5.341316709s" podCreationTimestamp="2025-12-01 21:41:20 +0000 UTC" firstStartedPulling="2025-12-01 21:41:22.225104894 +0000 UTC m=+400.715167241" lastFinishedPulling="2025-12-01 21:41:24.738729144 +0000 UTC m=+403.228791461" observedRunningTime="2025-12-01 21:41:25.335825858 +0000 UTC m=+403.825888185" watchObservedRunningTime="2025-12-01 21:41:25.341316709 +0000 UTC m=+403.831379026" Dec 01 21:41:26 crc kubenswrapper[4857]: I1201 21:41:26.351308 4857 generic.go:334] "Generic (PLEG): container finished" podID="c3e413ef-6fbc-4e78-8650-357ea026b79c" containerID="136e8a2a11e28a131644fa0b7b636f7eabd24e06d37059d5c345547105f95362" exitCode=0 Dec 01 21:41:26 crc kubenswrapper[4857]: I1201 21:41:26.351502 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8qfd" event={"ID":"c3e413ef-6fbc-4e78-8650-357ea026b79c","Type":"ContainerDied","Data":"136e8a2a11e28a131644fa0b7b636f7eabd24e06d37059d5c345547105f95362"} Dec 01 21:41:26 crc kubenswrapper[4857]: I1201 21:41:26.359237 4857 generic.go:334] "Generic (PLEG): container finished" podID="00a902c7-b8d4-4f4d-bc21-875579a6f0d0" containerID="49c394364721b00469739a0b4a4f8e71a6171b8ea671b27e5cc0cacbc40d9691" exitCode=0 Dec 01 21:41:26 crc kubenswrapper[4857]: I1201 21:41:26.360883 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mw8ff" event={"ID":"00a902c7-b8d4-4f4d-bc21-875579a6f0d0","Type":"ContainerDied","Data":"49c394364721b00469739a0b4a4f8e71a6171b8ea671b27e5cc0cacbc40d9691"} Dec 01 21:41:27 crc kubenswrapper[4857]: I1201 21:41:27.806581 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:41:27 crc kubenswrapper[4857]: I1201 21:41:27.808363 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:41:28 crc kubenswrapper[4857]: I1201 21:41:28.383627 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8qfd" event={"ID":"c3e413ef-6fbc-4e78-8650-357ea026b79c","Type":"ContainerStarted","Data":"b732822e026eb7b32175f5b9446eb5f8c7ebc159b52dbe88498ff2231b38f1aa"} Dec 01 21:41:28 crc kubenswrapper[4857]: I1201 21:41:28.388733 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mw8ff" event={"ID":"00a902c7-b8d4-4f4d-bc21-875579a6f0d0","Type":"ContainerStarted","Data":"9858a149e8bf902b9be211f4494411506a9cf14ea07153a4dd31804bd88373e3"} Dec 01 21:41:28 crc kubenswrapper[4857]: I1201 21:41:28.401377 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d8qfd" podStartSLOduration=3.659155845 podStartE2EDuration="6.40135604s" podCreationTimestamp="2025-12-01 21:41:22 +0000 UTC" firstStartedPulling="2025-12-01 21:41:24.269778349 +0000 UTC m=+402.759840696" lastFinishedPulling="2025-12-01 21:41:27.011978574 +0000 UTC m=+405.502040891" observedRunningTime="2025-12-01 21:41:28.398467226 +0000 UTC m=+406.888529553" watchObservedRunningTime="2025-12-01 21:41:28.40135604 +0000 UTC m=+406.891418357" Dec 01 21:41:28 crc kubenswrapper[4857]: I1201 21:41:28.423748 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mw8ff" podStartSLOduration=2.762445316 podStartE2EDuration="5.423730065s" podCreationTimestamp="2025-12-01 21:41:23 +0000 UTC" firstStartedPulling="2025-12-01 21:41:24.24898131 +0000 UTC m=+402.739043627" lastFinishedPulling="2025-12-01 21:41:26.910266049 +0000 UTC m=+405.400328376" observedRunningTime="2025-12-01 21:41:28.42016441 +0000 UTC m=+406.910226727" watchObservedRunningTime="2025-12-01 21:41:28.423730065 +0000 UTC m=+406.913792382" Dec 01 21:41:30 crc kubenswrapper[4857]: I1201 21:41:30.932737 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:30 crc kubenswrapper[4857]: I1201 21:41:30.933154 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:31 crc kubenswrapper[4857]: I1201 21:41:31.018072 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:31 crc kubenswrapper[4857]: I1201 21:41:31.106340 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:31 crc kubenswrapper[4857]: I1201 21:41:31.106492 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:31 crc kubenswrapper[4857]: I1201 21:41:31.156268 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:31 crc kubenswrapper[4857]: I1201 21:41:31.464217 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zfmgr" Dec 01 21:41:31 crc kubenswrapper[4857]: I1201 21:41:31.476124 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2x8zg" Dec 01 21:41:33 crc kubenswrapper[4857]: I1201 21:41:33.362138 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:33 crc kubenswrapper[4857]: I1201 21:41:33.362542 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:33 crc kubenswrapper[4857]: I1201 21:41:33.415847 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:33 crc kubenswrapper[4857]: I1201 21:41:33.492499 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:33 crc kubenswrapper[4857]: I1201 21:41:33.492555 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:33 crc kubenswrapper[4857]: I1201 21:41:33.526264 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d8qfd" Dec 01 21:41:33 crc kubenswrapper[4857]: I1201 21:41:33.545512 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:34 crc kubenswrapper[4857]: I1201 21:41:34.525879 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mw8ff" Dec 01 21:41:57 crc kubenswrapper[4857]: I1201 21:41:57.806132 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:41:57 crc kubenswrapper[4857]: I1201 21:41:57.806872 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:41:57 crc kubenswrapper[4857]: I1201 21:41:57.806939 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:41:57 crc kubenswrapper[4857]: I1201 21:41:57.807862 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"91968eb84103df465c19e9a8db52e323c98787b9caff59512702928cd30a7bb1"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 21:41:57 crc kubenswrapper[4857]: I1201 21:41:57.807967 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://91968eb84103df465c19e9a8db52e323c98787b9caff59512702928cd30a7bb1" gracePeriod=600 Dec 01 21:41:58 crc kubenswrapper[4857]: I1201 21:41:58.670725 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="91968eb84103df465c19e9a8db52e323c98787b9caff59512702928cd30a7bb1" exitCode=0 Dec 01 21:41:58 crc kubenswrapper[4857]: I1201 21:41:58.670797 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"91968eb84103df465c19e9a8db52e323c98787b9caff59512702928cd30a7bb1"} Dec 01 21:41:58 crc kubenswrapper[4857]: I1201 21:41:58.670843 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"697c28f6400e4a41370ca0b263896ab5c011bd878a53e6f826180b346cdc9317"} Dec 01 21:41:58 crc kubenswrapper[4857]: I1201 21:41:58.670875 4857 scope.go:117] "RemoveContainer" containerID="051a42e0faeab75c254cec9d9c3e056a155b08827528595798f8b67a43354786" Dec 01 21:44:27 crc kubenswrapper[4857]: I1201 21:44:27.806315 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:44:27 crc kubenswrapper[4857]: I1201 21:44:27.806930 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:44:57 crc kubenswrapper[4857]: I1201 21:44:57.806116 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:44:57 crc kubenswrapper[4857]: I1201 21:44:57.806948 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.224548 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27"] Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.288995 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.291877 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.298216 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.303733 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27"] Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.490481 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76127279-fe59-49e5-82cb-8f2d2a8afbe2-config-volume\") pod \"collect-profiles-29410425-q7l27\" (UID: \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.490619 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb4jt\" (UniqueName: \"kubernetes.io/projected/76127279-fe59-49e5-82cb-8f2d2a8afbe2-kube-api-access-sb4jt\") pod \"collect-profiles-29410425-q7l27\" (UID: \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.491330 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76127279-fe59-49e5-82cb-8f2d2a8afbe2-secret-volume\") pod \"collect-profiles-29410425-q7l27\" (UID: \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.592683 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76127279-fe59-49e5-82cb-8f2d2a8afbe2-config-volume\") pod \"collect-profiles-29410425-q7l27\" (UID: \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.592764 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb4jt\" (UniqueName: \"kubernetes.io/projected/76127279-fe59-49e5-82cb-8f2d2a8afbe2-kube-api-access-sb4jt\") pod \"collect-profiles-29410425-q7l27\" (UID: \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.592845 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76127279-fe59-49e5-82cb-8f2d2a8afbe2-secret-volume\") pod \"collect-profiles-29410425-q7l27\" (UID: \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.594509 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76127279-fe59-49e5-82cb-8f2d2a8afbe2-config-volume\") pod \"collect-profiles-29410425-q7l27\" (UID: \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.601909 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76127279-fe59-49e5-82cb-8f2d2a8afbe2-secret-volume\") pod \"collect-profiles-29410425-q7l27\" (UID: \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.638139 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb4jt\" (UniqueName: \"kubernetes.io/projected/76127279-fe59-49e5-82cb-8f2d2a8afbe2-kube-api-access-sb4jt\") pod \"collect-profiles-29410425-q7l27\" (UID: \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:00 crc kubenswrapper[4857]: I1201 21:45:00.907740 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:01 crc kubenswrapper[4857]: I1201 21:45:01.180978 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27"] Dec 01 21:45:02 crc kubenswrapper[4857]: I1201 21:45:02.081999 4857 generic.go:334] "Generic (PLEG): container finished" podID="76127279-fe59-49e5-82cb-8f2d2a8afbe2" containerID="3ad7b82ddc864cae02daef847d2c379ecfcb6459c26da987117dbb1ee24fab69" exitCode=0 Dec 01 21:45:02 crc kubenswrapper[4857]: I1201 21:45:02.082147 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" event={"ID":"76127279-fe59-49e5-82cb-8f2d2a8afbe2","Type":"ContainerDied","Data":"3ad7b82ddc864cae02daef847d2c379ecfcb6459c26da987117dbb1ee24fab69"} Dec 01 21:45:02 crc kubenswrapper[4857]: I1201 21:45:02.083226 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" event={"ID":"76127279-fe59-49e5-82cb-8f2d2a8afbe2","Type":"ContainerStarted","Data":"911eb9d9b5340b41dc359de0ab6aa4339310d0ded286c870424505b040884cea"} Dec 01 21:45:03 crc kubenswrapper[4857]: I1201 21:45:03.372625 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:03 crc kubenswrapper[4857]: I1201 21:45:03.532936 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76127279-fe59-49e5-82cb-8f2d2a8afbe2-secret-volume\") pod \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\" (UID: \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\") " Dec 01 21:45:03 crc kubenswrapper[4857]: I1201 21:45:03.533181 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb4jt\" (UniqueName: \"kubernetes.io/projected/76127279-fe59-49e5-82cb-8f2d2a8afbe2-kube-api-access-sb4jt\") pod \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\" (UID: \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\") " Dec 01 21:45:03 crc kubenswrapper[4857]: I1201 21:45:03.533259 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76127279-fe59-49e5-82cb-8f2d2a8afbe2-config-volume\") pod \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\" (UID: \"76127279-fe59-49e5-82cb-8f2d2a8afbe2\") " Dec 01 21:45:03 crc kubenswrapper[4857]: I1201 21:45:03.534434 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76127279-fe59-49e5-82cb-8f2d2a8afbe2-config-volume" (OuterVolumeSpecName: "config-volume") pod "76127279-fe59-49e5-82cb-8f2d2a8afbe2" (UID: "76127279-fe59-49e5-82cb-8f2d2a8afbe2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:45:03 crc kubenswrapper[4857]: I1201 21:45:03.543261 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76127279-fe59-49e5-82cb-8f2d2a8afbe2-kube-api-access-sb4jt" (OuterVolumeSpecName: "kube-api-access-sb4jt") pod "76127279-fe59-49e5-82cb-8f2d2a8afbe2" (UID: "76127279-fe59-49e5-82cb-8f2d2a8afbe2"). InnerVolumeSpecName "kube-api-access-sb4jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:45:03 crc kubenswrapper[4857]: I1201 21:45:03.543272 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76127279-fe59-49e5-82cb-8f2d2a8afbe2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "76127279-fe59-49e5-82cb-8f2d2a8afbe2" (UID: "76127279-fe59-49e5-82cb-8f2d2a8afbe2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:45:03 crc kubenswrapper[4857]: I1201 21:45:03.635371 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb4jt\" (UniqueName: \"kubernetes.io/projected/76127279-fe59-49e5-82cb-8f2d2a8afbe2-kube-api-access-sb4jt\") on node \"crc\" DevicePath \"\"" Dec 01 21:45:03 crc kubenswrapper[4857]: I1201 21:45:03.635426 4857 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76127279-fe59-49e5-82cb-8f2d2a8afbe2-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 21:45:03 crc kubenswrapper[4857]: I1201 21:45:03.635445 4857 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76127279-fe59-49e5-82cb-8f2d2a8afbe2-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 21:45:04 crc kubenswrapper[4857]: I1201 21:45:04.101140 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" Dec 01 21:45:04 crc kubenswrapper[4857]: I1201 21:45:04.101551 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27" event={"ID":"76127279-fe59-49e5-82cb-8f2d2a8afbe2","Type":"ContainerDied","Data":"911eb9d9b5340b41dc359de0ab6aa4339310d0ded286c870424505b040884cea"} Dec 01 21:45:04 crc kubenswrapper[4857]: I1201 21:45:04.101611 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="911eb9d9b5340b41dc359de0ab6aa4339310d0ded286c870424505b040884cea" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.779015 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nzh9t"] Dec 01 21:45:06 crc kubenswrapper[4857]: E1201 21:45:06.779757 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76127279-fe59-49e5-82cb-8f2d2a8afbe2" containerName="collect-profiles" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.779779 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="76127279-fe59-49e5-82cb-8f2d2a8afbe2" containerName="collect-profiles" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.779982 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="76127279-fe59-49e5-82cb-8f2d2a8afbe2" containerName="collect-profiles" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.780587 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.805399 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nzh9t"] Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.882296 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m9wr\" (UniqueName: \"kubernetes.io/projected/da87040f-3ebf-4ab9-bff6-fa1b3d311482-kube-api-access-7m9wr\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.882379 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/da87040f-3ebf-4ab9-bff6-fa1b3d311482-registry-certificates\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.882460 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/da87040f-3ebf-4ab9-bff6-fa1b3d311482-registry-tls\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.882484 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/da87040f-3ebf-4ab9-bff6-fa1b3d311482-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.882696 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/da87040f-3ebf-4ab9-bff6-fa1b3d311482-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.882782 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/da87040f-3ebf-4ab9-bff6-fa1b3d311482-trusted-ca\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.882803 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/da87040f-3ebf-4ab9-bff6-fa1b3d311482-bound-sa-token\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.882856 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.910898 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.984437 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/da87040f-3ebf-4ab9-bff6-fa1b3d311482-registry-tls\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.984504 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/da87040f-3ebf-4ab9-bff6-fa1b3d311482-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.984572 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/da87040f-3ebf-4ab9-bff6-fa1b3d311482-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.984624 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/da87040f-3ebf-4ab9-bff6-fa1b3d311482-trusted-ca\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.984654 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/da87040f-3ebf-4ab9-bff6-fa1b3d311482-bound-sa-token\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.984702 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m9wr\" (UniqueName: \"kubernetes.io/projected/da87040f-3ebf-4ab9-bff6-fa1b3d311482-kube-api-access-7m9wr\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.984743 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/da87040f-3ebf-4ab9-bff6-fa1b3d311482-registry-certificates\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.985361 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/da87040f-3ebf-4ab9-bff6-fa1b3d311482-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.986692 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/da87040f-3ebf-4ab9-bff6-fa1b3d311482-registry-certificates\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.987600 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/da87040f-3ebf-4ab9-bff6-fa1b3d311482-trusted-ca\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.990985 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/da87040f-3ebf-4ab9-bff6-fa1b3d311482-registry-tls\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:06 crc kubenswrapper[4857]: I1201 21:45:06.999738 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/da87040f-3ebf-4ab9-bff6-fa1b3d311482-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:07 crc kubenswrapper[4857]: I1201 21:45:07.006563 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m9wr\" (UniqueName: \"kubernetes.io/projected/da87040f-3ebf-4ab9-bff6-fa1b3d311482-kube-api-access-7m9wr\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:07 crc kubenswrapper[4857]: I1201 21:45:07.013581 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/da87040f-3ebf-4ab9-bff6-fa1b3d311482-bound-sa-token\") pod \"image-registry-66df7c8f76-nzh9t\" (UID: \"da87040f-3ebf-4ab9-bff6-fa1b3d311482\") " pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:07 crc kubenswrapper[4857]: I1201 21:45:07.095361 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:07 crc kubenswrapper[4857]: I1201 21:45:07.416870 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nzh9t"] Dec 01 21:45:07 crc kubenswrapper[4857]: W1201 21:45:07.423077 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda87040f_3ebf_4ab9_bff6_fa1b3d311482.slice/crio-f8c6d9ef552fb9726f8e449280a8577d03d74257ee1115ed0f1642f6d2366c0c WatchSource:0}: Error finding container f8c6d9ef552fb9726f8e449280a8577d03d74257ee1115ed0f1642f6d2366c0c: Status 404 returned error can't find the container with id f8c6d9ef552fb9726f8e449280a8577d03d74257ee1115ed0f1642f6d2366c0c Dec 01 21:45:08 crc kubenswrapper[4857]: I1201 21:45:08.128017 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" event={"ID":"da87040f-3ebf-4ab9-bff6-fa1b3d311482","Type":"ContainerStarted","Data":"8192bd1706544f1f1c2358e5dd3dc59058e14275eceadbb22e6923b50840b2cb"} Dec 01 21:45:08 crc kubenswrapper[4857]: I1201 21:45:08.128103 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" event={"ID":"da87040f-3ebf-4ab9-bff6-fa1b3d311482","Type":"ContainerStarted","Data":"f8c6d9ef552fb9726f8e449280a8577d03d74257ee1115ed0f1642f6d2366c0c"} Dec 01 21:45:08 crc kubenswrapper[4857]: I1201 21:45:08.128226 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:27 crc kubenswrapper[4857]: I1201 21:45:27.103304 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" Dec 01 21:45:27 crc kubenswrapper[4857]: I1201 21:45:27.192743 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-nzh9t" podStartSLOduration=21.19271833 podStartE2EDuration="21.19271833s" podCreationTimestamp="2025-12-01 21:45:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:45:08.154215881 +0000 UTC m=+626.644278218" watchObservedRunningTime="2025-12-01 21:45:27.19271833 +0000 UTC m=+645.682780667" Dec 01 21:45:27 crc kubenswrapper[4857]: I1201 21:45:27.221583 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zdzg5"] Dec 01 21:45:27 crc kubenswrapper[4857]: I1201 21:45:27.806185 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:45:27 crc kubenswrapper[4857]: I1201 21:45:27.806755 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:45:27 crc kubenswrapper[4857]: I1201 21:45:27.806849 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:45:27 crc kubenswrapper[4857]: I1201 21:45:27.808001 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"697c28f6400e4a41370ca0b263896ab5c011bd878a53e6f826180b346cdc9317"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 21:45:27 crc kubenswrapper[4857]: I1201 21:45:27.808296 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://697c28f6400e4a41370ca0b263896ab5c011bd878a53e6f826180b346cdc9317" gracePeriod=600 Dec 01 21:45:28 crc kubenswrapper[4857]: I1201 21:45:28.273937 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="697c28f6400e4a41370ca0b263896ab5c011bd878a53e6f826180b346cdc9317" exitCode=0 Dec 01 21:45:28 crc kubenswrapper[4857]: I1201 21:45:28.274000 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"697c28f6400e4a41370ca0b263896ab5c011bd878a53e6f826180b346cdc9317"} Dec 01 21:45:28 crc kubenswrapper[4857]: I1201 21:45:28.274070 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"f890539e4830c07b0e2e556ca8ce83ae288e8de08c435fa380b5373247a6c2d8"} Dec 01 21:45:28 crc kubenswrapper[4857]: I1201 21:45:28.274099 4857 scope.go:117] "RemoveContainer" containerID="91968eb84103df465c19e9a8db52e323c98787b9caff59512702928cd30a7bb1" Dec 01 21:45:42 crc kubenswrapper[4857]: I1201 21:45:42.127286 4857 scope.go:117] "RemoveContainer" containerID="db076f2683fbbb30d0b00304714478cf038f8a8272c27b3bec2b83b75bd8c928" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.311679 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" podUID="658992cc-be88-49f4-b35c-bc445583fc39" containerName="registry" containerID="cri-o://28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e" gracePeriod=30 Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.763686 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.829200 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-bound-sa-token\") pod \"658992cc-be88-49f4-b35c-bc445583fc39\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.829333 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp9pd\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-kube-api-access-zp9pd\") pod \"658992cc-be88-49f4-b35c-bc445583fc39\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.829405 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/658992cc-be88-49f4-b35c-bc445583fc39-trusted-ca\") pod \"658992cc-be88-49f4-b35c-bc445583fc39\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.829445 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/658992cc-be88-49f4-b35c-bc445583fc39-installation-pull-secrets\") pod \"658992cc-be88-49f4-b35c-bc445583fc39\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.829476 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/658992cc-be88-49f4-b35c-bc445583fc39-registry-certificates\") pod \"658992cc-be88-49f4-b35c-bc445583fc39\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.829557 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/658992cc-be88-49f4-b35c-bc445583fc39-ca-trust-extracted\") pod \"658992cc-be88-49f4-b35c-bc445583fc39\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.829806 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"658992cc-be88-49f4-b35c-bc445583fc39\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.829845 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-registry-tls\") pod \"658992cc-be88-49f4-b35c-bc445583fc39\" (UID: \"658992cc-be88-49f4-b35c-bc445583fc39\") " Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.831384 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/658992cc-be88-49f4-b35c-bc445583fc39-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "658992cc-be88-49f4-b35c-bc445583fc39" (UID: "658992cc-be88-49f4-b35c-bc445583fc39"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.831889 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/658992cc-be88-49f4-b35c-bc445583fc39-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "658992cc-be88-49f4-b35c-bc445583fc39" (UID: "658992cc-be88-49f4-b35c-bc445583fc39"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.839142 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/658992cc-be88-49f4-b35c-bc445583fc39-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "658992cc-be88-49f4-b35c-bc445583fc39" (UID: "658992cc-be88-49f4-b35c-bc445583fc39"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.839546 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-kube-api-access-zp9pd" (OuterVolumeSpecName: "kube-api-access-zp9pd") pod "658992cc-be88-49f4-b35c-bc445583fc39" (UID: "658992cc-be88-49f4-b35c-bc445583fc39"). InnerVolumeSpecName "kube-api-access-zp9pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.844995 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "658992cc-be88-49f4-b35c-bc445583fc39" (UID: "658992cc-be88-49f4-b35c-bc445583fc39"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.845953 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "658992cc-be88-49f4-b35c-bc445583fc39" (UID: "658992cc-be88-49f4-b35c-bc445583fc39"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.847638 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "658992cc-be88-49f4-b35c-bc445583fc39" (UID: "658992cc-be88-49f4-b35c-bc445583fc39"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.864980 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/658992cc-be88-49f4-b35c-bc445583fc39-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "658992cc-be88-49f4-b35c-bc445583fc39" (UID: "658992cc-be88-49f4-b35c-bc445583fc39"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.932363 4857 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/658992cc-be88-49f4-b35c-bc445583fc39-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.932422 4857 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.932461 4857 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.932490 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp9pd\" (UniqueName: \"kubernetes.io/projected/658992cc-be88-49f4-b35c-bc445583fc39-kube-api-access-zp9pd\") on node \"crc\" DevicePath \"\"" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.932518 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/658992cc-be88-49f4-b35c-bc445583fc39-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.932540 4857 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/658992cc-be88-49f4-b35c-bc445583fc39-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 21:45:52 crc kubenswrapper[4857]: I1201 21:45:52.932566 4857 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/658992cc-be88-49f4-b35c-bc445583fc39-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 21:45:53 crc kubenswrapper[4857]: I1201 21:45:53.456254 4857 generic.go:334] "Generic (PLEG): container finished" podID="658992cc-be88-49f4-b35c-bc445583fc39" containerID="28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e" exitCode=0 Dec 01 21:45:53 crc kubenswrapper[4857]: I1201 21:45:53.456422 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" Dec 01 21:45:53 crc kubenswrapper[4857]: I1201 21:45:53.456461 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" event={"ID":"658992cc-be88-49f4-b35c-bc445583fc39","Type":"ContainerDied","Data":"28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e"} Dec 01 21:45:53 crc kubenswrapper[4857]: I1201 21:45:53.457546 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zdzg5" event={"ID":"658992cc-be88-49f4-b35c-bc445583fc39","Type":"ContainerDied","Data":"2b469c820b65d666f421b57cbc052f8336afa928a6a52f2aeeba9f525091e944"} Dec 01 21:45:53 crc kubenswrapper[4857]: I1201 21:45:53.457614 4857 scope.go:117] "RemoveContainer" containerID="28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e" Dec 01 21:45:53 crc kubenswrapper[4857]: I1201 21:45:53.497261 4857 scope.go:117] "RemoveContainer" containerID="28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e" Dec 01 21:45:53 crc kubenswrapper[4857]: E1201 21:45:53.498489 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e\": container with ID starting with 28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e not found: ID does not exist" containerID="28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e" Dec 01 21:45:53 crc kubenswrapper[4857]: I1201 21:45:53.498574 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e"} err="failed to get container status \"28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e\": rpc error: code = NotFound desc = could not find container \"28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e\": container with ID starting with 28b1e4d70777ccd474142e3a4ad2782d3b0fcf27728ad15c9b03db7de7e6c41e not found: ID does not exist" Dec 01 21:45:53 crc kubenswrapper[4857]: I1201 21:45:53.583118 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zdzg5"] Dec 01 21:45:53 crc kubenswrapper[4857]: I1201 21:45:53.586154 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zdzg5"] Dec 01 21:45:53 crc kubenswrapper[4857]: I1201 21:45:53.847621 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="658992cc-be88-49f4-b35c-bc445583fc39" path="/var/lib/kubelet/pods/658992cc-be88-49f4-b35c-bc445583fc39/volumes" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.578416 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-lcww8"] Dec 01 21:46:36 crc kubenswrapper[4857]: E1201 21:46:36.579279 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="658992cc-be88-49f4-b35c-bc445583fc39" containerName="registry" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.579295 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="658992cc-be88-49f4-b35c-bc445583fc39" containerName="registry" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.579442 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="658992cc-be88-49f4-b35c-bc445583fc39" containerName="registry" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.579863 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-lcww8" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.582131 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.582424 4857 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-z7jq9" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.582817 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.596257 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-lcww8"] Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.616671 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-wnq4k"] Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.618476 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-wnq4k" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.634302 4857 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-ng62h" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.651114 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-5gv57"] Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.651860 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-5gv57" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.653922 4857 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-bt5lq" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.656471 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-5gv57"] Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.661225 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-wnq4k"] Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.734539 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvxpm\" (UniqueName: \"kubernetes.io/projected/84a659be-d476-4ba1-820e-c5279704aec5-kube-api-access-vvxpm\") pod \"cert-manager-cainjector-7f985d654d-lcww8\" (UID: \"84a659be-d476-4ba1-820e-c5279704aec5\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-lcww8" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.734624 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2qf2\" (UniqueName: \"kubernetes.io/projected/e8804d0a-50d0-4b87-bc77-a8c8229ccf1d-kube-api-access-b2qf2\") pod \"cert-manager-5b446d88c5-wnq4k\" (UID: \"e8804d0a-50d0-4b87-bc77-a8c8229ccf1d\") " pod="cert-manager/cert-manager-5b446d88c5-wnq4k" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.835716 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2qf2\" (UniqueName: \"kubernetes.io/projected/e8804d0a-50d0-4b87-bc77-a8c8229ccf1d-kube-api-access-b2qf2\") pod \"cert-manager-5b446d88c5-wnq4k\" (UID: \"e8804d0a-50d0-4b87-bc77-a8c8229ccf1d\") " pod="cert-manager/cert-manager-5b446d88c5-wnq4k" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.835844 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvwhz\" (UniqueName: \"kubernetes.io/projected/64fd8572-2a45-422e-a413-49314dc52a55-kube-api-access-dvwhz\") pod \"cert-manager-webhook-5655c58dd6-5gv57\" (UID: \"64fd8572-2a45-422e-a413-49314dc52a55\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-5gv57" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.835896 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvxpm\" (UniqueName: \"kubernetes.io/projected/84a659be-d476-4ba1-820e-c5279704aec5-kube-api-access-vvxpm\") pod \"cert-manager-cainjector-7f985d654d-lcww8\" (UID: \"84a659be-d476-4ba1-820e-c5279704aec5\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-lcww8" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.858320 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvxpm\" (UniqueName: \"kubernetes.io/projected/84a659be-d476-4ba1-820e-c5279704aec5-kube-api-access-vvxpm\") pod \"cert-manager-cainjector-7f985d654d-lcww8\" (UID: \"84a659be-d476-4ba1-820e-c5279704aec5\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-lcww8" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.859433 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2qf2\" (UniqueName: \"kubernetes.io/projected/e8804d0a-50d0-4b87-bc77-a8c8229ccf1d-kube-api-access-b2qf2\") pod \"cert-manager-5b446d88c5-wnq4k\" (UID: \"e8804d0a-50d0-4b87-bc77-a8c8229ccf1d\") " pod="cert-manager/cert-manager-5b446d88c5-wnq4k" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.907706 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-lcww8" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.936913 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvwhz\" (UniqueName: \"kubernetes.io/projected/64fd8572-2a45-422e-a413-49314dc52a55-kube-api-access-dvwhz\") pod \"cert-manager-webhook-5655c58dd6-5gv57\" (UID: \"64fd8572-2a45-422e-a413-49314dc52a55\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-5gv57" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.948236 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-wnq4k" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.959015 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvwhz\" (UniqueName: \"kubernetes.io/projected/64fd8572-2a45-422e-a413-49314dc52a55-kube-api-access-dvwhz\") pod \"cert-manager-webhook-5655c58dd6-5gv57\" (UID: \"64fd8572-2a45-422e-a413-49314dc52a55\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-5gv57" Dec 01 21:46:36 crc kubenswrapper[4857]: I1201 21:46:36.964218 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-5gv57" Dec 01 21:46:37 crc kubenswrapper[4857]: I1201 21:46:37.177452 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-lcww8"] Dec 01 21:46:37 crc kubenswrapper[4857]: W1201 21:46:37.182482 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84a659be_d476_4ba1_820e_c5279704aec5.slice/crio-0f80dd0f6d97dbee6ccf5418a8789735aaaa4b3c008e7278117edfd00300523d WatchSource:0}: Error finding container 0f80dd0f6d97dbee6ccf5418a8789735aaaa4b3c008e7278117edfd00300523d: Status 404 returned error can't find the container with id 0f80dd0f6d97dbee6ccf5418a8789735aaaa4b3c008e7278117edfd00300523d Dec 01 21:46:37 crc kubenswrapper[4857]: I1201 21:46:37.188090 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 21:46:37 crc kubenswrapper[4857]: I1201 21:46:37.218719 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-5gv57"] Dec 01 21:46:37 crc kubenswrapper[4857]: W1201 21:46:37.222150 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64fd8572_2a45_422e_a413_49314dc52a55.slice/crio-7c7e7c0c807618e43cee60e7b5f09e5391c682265e83d5863aa2025496ae971d WatchSource:0}: Error finding container 7c7e7c0c807618e43cee60e7b5f09e5391c682265e83d5863aa2025496ae971d: Status 404 returned error can't find the container with id 7c7e7c0c807618e43cee60e7b5f09e5391c682265e83d5863aa2025496ae971d Dec 01 21:46:37 crc kubenswrapper[4857]: I1201 21:46:37.263337 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-wnq4k"] Dec 01 21:46:37 crc kubenswrapper[4857]: W1201 21:46:37.270947 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8804d0a_50d0_4b87_bc77_a8c8229ccf1d.slice/crio-a5d47aeffa8eee48d362da1f0e2b83a575b9b7dbf5f19be556f7fefb31a36429 WatchSource:0}: Error finding container a5d47aeffa8eee48d362da1f0e2b83a575b9b7dbf5f19be556f7fefb31a36429: Status 404 returned error can't find the container with id a5d47aeffa8eee48d362da1f0e2b83a575b9b7dbf5f19be556f7fefb31a36429 Dec 01 21:46:37 crc kubenswrapper[4857]: I1201 21:46:37.756602 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-wnq4k" event={"ID":"e8804d0a-50d0-4b87-bc77-a8c8229ccf1d","Type":"ContainerStarted","Data":"a5d47aeffa8eee48d362da1f0e2b83a575b9b7dbf5f19be556f7fefb31a36429"} Dec 01 21:46:37 crc kubenswrapper[4857]: I1201 21:46:37.759866 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-lcww8" event={"ID":"84a659be-d476-4ba1-820e-c5279704aec5","Type":"ContainerStarted","Data":"0f80dd0f6d97dbee6ccf5418a8789735aaaa4b3c008e7278117edfd00300523d"} Dec 01 21:46:37 crc kubenswrapper[4857]: I1201 21:46:37.762105 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-5gv57" event={"ID":"64fd8572-2a45-422e-a413-49314dc52a55","Type":"ContainerStarted","Data":"7c7e7c0c807618e43cee60e7b5f09e5391c682265e83d5863aa2025496ae971d"} Dec 01 21:46:40 crc kubenswrapper[4857]: I1201 21:46:40.784930 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-lcww8" event={"ID":"84a659be-d476-4ba1-820e-c5279704aec5","Type":"ContainerStarted","Data":"d359f12c27960a259d093a40f8fbabd89ff59b63c2f5f3e961025789cd15e9a6"} Dec 01 21:46:40 crc kubenswrapper[4857]: I1201 21:46:40.789109 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-5gv57" event={"ID":"64fd8572-2a45-422e-a413-49314dc52a55","Type":"ContainerStarted","Data":"79f12a7e711f1ab26d894dd724ea02f84c13ddf3aa2659737319eaf1412c4452"} Dec 01 21:46:40 crc kubenswrapper[4857]: I1201 21:46:40.789475 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-5gv57" Dec 01 21:46:40 crc kubenswrapper[4857]: I1201 21:46:40.803062 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-lcww8" podStartSLOduration=1.995407757 podStartE2EDuration="4.803022789s" podCreationTimestamp="2025-12-01 21:46:36 +0000 UTC" firstStartedPulling="2025-12-01 21:46:37.18779634 +0000 UTC m=+715.677858647" lastFinishedPulling="2025-12-01 21:46:39.995411362 +0000 UTC m=+718.485473679" observedRunningTime="2025-12-01 21:46:40.802447545 +0000 UTC m=+719.292509872" watchObservedRunningTime="2025-12-01 21:46:40.803022789 +0000 UTC m=+719.293085106" Dec 01 21:46:40 crc kubenswrapper[4857]: I1201 21:46:40.824789 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-5gv57" podStartSLOduration=2.107439158 podStartE2EDuration="4.824755871s" podCreationTimestamp="2025-12-01 21:46:36 +0000 UTC" firstStartedPulling="2025-12-01 21:46:37.225168901 +0000 UTC m=+715.715231218" lastFinishedPulling="2025-12-01 21:46:39.942485614 +0000 UTC m=+718.432547931" observedRunningTime="2025-12-01 21:46:40.818623588 +0000 UTC m=+719.308685915" watchObservedRunningTime="2025-12-01 21:46:40.824755871 +0000 UTC m=+719.314818188" Dec 01 21:46:41 crc kubenswrapper[4857]: I1201 21:46:41.799109 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-wnq4k" event={"ID":"e8804d0a-50d0-4b87-bc77-a8c8229ccf1d","Type":"ContainerStarted","Data":"f3fa14f5f4af46015cb314f912c66f7b2e1e015c18f508fb237ed179b7da514e"} Dec 01 21:46:41 crc kubenswrapper[4857]: I1201 21:46:41.826230 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-wnq4k" podStartSLOduration=2.063624407 podStartE2EDuration="5.826196794s" podCreationTimestamp="2025-12-01 21:46:36 +0000 UTC" firstStartedPulling="2025-12-01 21:46:37.273878094 +0000 UTC m=+715.763940421" lastFinishedPulling="2025-12-01 21:46:41.036450471 +0000 UTC m=+719.526512808" observedRunningTime="2025-12-01 21:46:41.816875232 +0000 UTC m=+720.306937579" watchObservedRunningTime="2025-12-01 21:46:41.826196794 +0000 UTC m=+720.316259151" Dec 01 21:46:46 crc kubenswrapper[4857]: I1201 21:46:46.968148 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-5gv57" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.300428 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mdc5b"] Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.301626 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovn-controller" containerID="cri-o://dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d" gracePeriod=30 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.302172 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="northd" containerID="cri-o://825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462" gracePeriod=30 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.302327 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723" gracePeriod=30 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.302427 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="kube-rbac-proxy-node" containerID="cri-o://17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe" gracePeriod=30 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.302511 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovn-acl-logging" containerID="cri-o://8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078" gracePeriod=30 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.302733 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="nbdb" containerID="cri-o://49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3" gracePeriod=30 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.302853 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="sbdb" containerID="cri-o://49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c" gracePeriod=30 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.352668 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" containerID="cri-o://1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca" gracePeriod=30 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.615407 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/3.log" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.617887 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovn-acl-logging/0.log" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.618504 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovn-controller/0.log" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.619027 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.675853 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-69n7v"] Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676132 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676147 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676157 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676166 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676176 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="kubecfg-setup" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676184 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="kubecfg-setup" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676195 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovn-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676204 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovn-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676217 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="sbdb" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676225 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="sbdb" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676237 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676245 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676255 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="kube-rbac-proxy-node" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676262 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="kube-rbac-proxy-node" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676272 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="nbdb" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676280 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="nbdb" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676295 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovn-acl-logging" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676303 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovn-acl-logging" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676316 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="northd" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676324 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="northd" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676337 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676345 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676459 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovn-acl-logging" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676474 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676484 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="northd" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676494 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676503 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676513 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="sbdb" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676523 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovn-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676531 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="nbdb" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676544 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="kube-rbac-proxy-node" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676554 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676671 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676680 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.676695 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676703 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.676831 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.677067 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerName="ovnkube-controller" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.678932 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.814717 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-cni-netd\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.814793 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovnkube-config\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.814824 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-openvswitch\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.814864 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-run-ovn-kubernetes\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.814862 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.814901 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-slash\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.814955 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-slash" (OuterVolumeSpecName: "host-slash") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815003 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-log-socket\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815010 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815073 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-log-socket" (OuterVolumeSpecName: "log-socket") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815085 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-run-netns\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815024 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815110 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815132 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-systemd-units\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815178 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-ovn\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815211 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovn-node-metrics-cert\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815238 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-cni-bin\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815240 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815274 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovnkube-script-lib\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815311 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-etc-openvswitch\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815297 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815358 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815332 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-kubelet\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815389 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815452 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-var-lib-cni-networks-ovn-kubernetes\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815462 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815505 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-node-log\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815523 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815543 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-var-lib-openvswitch\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815562 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-node-log" (OuterVolumeSpecName: "node-log") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815604 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-env-overrides\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815649 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-systemd\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815682 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftp2p\" (UniqueName: \"kubernetes.io/projected/557fd3cd-9ab9-4135-af14-f748fdc34a90-kube-api-access-ftp2p\") pod \"557fd3cd-9ab9-4135-af14-f748fdc34a90\" (UID: \"557fd3cd-9ab9-4135-af14-f748fdc34a90\") " Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815683 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815924 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-cni-bin\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.815962 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-node-log\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816023 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn99b\" (UniqueName: \"kubernetes.io/projected/1076c25c-e534-4fb3-8b76-1e36bd1897bd-kube-api-access-rn99b\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816138 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-systemd-units\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816199 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-run-openvswitch\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816557 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-etc-openvswitch\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816596 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1076c25c-e534-4fb3-8b76-1e36bd1897bd-ovnkube-config\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816274 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816440 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816655 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-run-ovn\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816757 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-kubelet\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816787 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-run-systemd\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816810 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1076c25c-e534-4fb3-8b76-1e36bd1897bd-env-overrides\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816835 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1076c25c-e534-4fb3-8b76-1e36bd1897bd-ovnkube-script-lib\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816876 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.816985 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-slash\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817140 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-log-socket\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817219 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-run-ovn-kubernetes\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817444 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1076c25c-e534-4fb3-8b76-1e36bd1897bd-ovn-node-metrics-cert\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817492 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-cni-netd\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817544 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-run-netns\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817584 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-var-lib-openvswitch\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817768 4857 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817811 4857 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-slash\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817816 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817839 4857 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-log-socket\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817907 4857 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817934 4857 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817963 4857 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.817989 4857 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.818014 4857 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.818073 4857 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.818100 4857 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.818125 4857 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.818151 4857 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-node-log\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.818173 4857 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.818193 4857 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.818235 4857 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.818253 4857 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.825120 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.825682 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/557fd3cd-9ab9-4135-af14-f748fdc34a90-kube-api-access-ftp2p" (OuterVolumeSpecName: "kube-api-access-ftp2p") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "kube-api-access-ftp2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.829848 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "557fd3cd-9ab9-4135-af14-f748fdc34a90" (UID: "557fd3cd-9ab9-4135-af14-f748fdc34a90"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.840979 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dfxtl_fc7d55e5-8085-40b7-8147-8508ee6c3f61/kube-multus/2.log" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.841613 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dfxtl_fc7d55e5-8085-40b7-8147-8508ee6c3f61/kube-multus/1.log" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.841648 4857 generic.go:334] "Generic (PLEG): container finished" podID="fc7d55e5-8085-40b7-8147-8508ee6c3f61" containerID="0d75201bf392092adaf26780aa1582bd7f68dae7e1f0c99d67785a27dd80b8f8" exitCode=2 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.844243 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovnkube-controller/3.log" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.845571 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dfxtl" event={"ID":"fc7d55e5-8085-40b7-8147-8508ee6c3f61","Type":"ContainerDied","Data":"0d75201bf392092adaf26780aa1582bd7f68dae7e1f0c99d67785a27dd80b8f8"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.845613 4857 scope.go:117] "RemoveContainer" containerID="6e38c5755141156082b254a121520d58f44c5cef43ad63c3cadb2f1cf8e186e7" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.846489 4857 scope.go:117] "RemoveContainer" containerID="0d75201bf392092adaf26780aa1582bd7f68dae7e1f0c99d67785a27dd80b8f8" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.847639 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovn-acl-logging/0.log" Dec 01 21:46:47 crc kubenswrapper[4857]: E1201 21:46:47.848214 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-dfxtl_openshift-multus(fc7d55e5-8085-40b7-8147-8508ee6c3f61)\"" pod="openshift-multus/multus-dfxtl" podUID="fc7d55e5-8085-40b7-8147-8508ee6c3f61" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.849396 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mdc5b_557fd3cd-9ab9-4135-af14-f748fdc34a90/ovn-controller/0.log" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.850323 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca" exitCode=0 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.850400 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c" exitCode=0 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.850460 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3" exitCode=0 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.850476 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462" exitCode=0 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.850488 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723" exitCode=0 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.850500 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe" exitCode=0 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.850577 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078" exitCode=143 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.850595 4857 generic.go:334] "Generic (PLEG): container finished" podID="557fd3cd-9ab9-4135-af14-f748fdc34a90" containerID="dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d" exitCode=143 Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.850630 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.850715 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851014 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851138 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851224 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851291 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851318 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851336 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851385 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851397 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851408 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851419 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851466 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851478 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851489 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851499 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851515 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851567 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851583 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851594 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851604 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851651 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851664 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851880 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851892 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851903 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851913 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851099 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.851964 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852089 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852107 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852117 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852124 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852131 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852138 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852146 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852152 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852159 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852165 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852177 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mdc5b" event={"ID":"557fd3cd-9ab9-4135-af14-f748fdc34a90","Type":"ContainerDied","Data":"b7de8baf745835264f416445383bebf888433f4ef55305b0795912a2e4619616"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852193 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852202 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852209 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852216 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852223 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852230 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852237 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852245 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852252 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.852260 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f"} Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.879906 4857 scope.go:117] "RemoveContainer" containerID="1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.895020 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mdc5b"] Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.901836 4857 scope.go:117] "RemoveContainer" containerID="9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.910066 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mdc5b"] Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919265 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-run-openvswitch\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919304 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-etc-openvswitch\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919327 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1076c25c-e534-4fb3-8b76-1e36bd1897bd-ovnkube-config\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919352 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-run-ovn\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919368 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-run-systemd\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919384 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-kubelet\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919408 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1076c25c-e534-4fb3-8b76-1e36bd1897bd-env-overrides\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919450 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1076c25c-e534-4fb3-8b76-1e36bd1897bd-ovnkube-script-lib\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919474 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919496 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-slash\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919503 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-run-ovn\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919521 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-log-socket\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919527 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-run-openvswitch\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919473 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-run-systemd\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919579 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-slash\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919580 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919600 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-run-ovn-kubernetes\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919561 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-log-socket\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919631 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-run-ovn-kubernetes\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919633 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-etc-openvswitch\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919648 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-kubelet\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919733 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1076c25c-e534-4fb3-8b76-1e36bd1897bd-ovn-node-metrics-cert\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919763 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-cni-netd\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919798 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-run-netns\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919819 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-var-lib-openvswitch\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919846 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-cni-bin\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919872 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-node-log\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919883 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-cni-netd\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919927 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-var-lib-openvswitch\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.919934 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn99b\" (UniqueName: \"kubernetes.io/projected/1076c25c-e534-4fb3-8b76-1e36bd1897bd-kube-api-access-rn99b\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.920014 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-systemd-units\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.920115 4857 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/557fd3cd-9ab9-4135-af14-f748fdc34a90-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.920131 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftp2p\" (UniqueName: \"kubernetes.io/projected/557fd3cd-9ab9-4135-af14-f748fdc34a90-kube-api-access-ftp2p\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.920147 4857 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.920162 4857 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/557fd3cd-9ab9-4135-af14-f748fdc34a90-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.920195 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-systemd-units\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.920230 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-run-netns\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.920239 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-host-cni-bin\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.920252 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1076c25c-e534-4fb3-8b76-1e36bd1897bd-env-overrides\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.920271 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1076c25c-e534-4fb3-8b76-1e36bd1897bd-node-log\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.921187 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1076c25c-e534-4fb3-8b76-1e36bd1897bd-ovnkube-config\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.921247 4857 scope.go:117] "RemoveContainer" containerID="49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.921344 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1076c25c-e534-4fb3-8b76-1e36bd1897bd-ovnkube-script-lib\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.923788 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1076c25c-e534-4fb3-8b76-1e36bd1897bd-ovn-node-metrics-cert\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.936013 4857 scope.go:117] "RemoveContainer" containerID="49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.940470 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn99b\" (UniqueName: \"kubernetes.io/projected/1076c25c-e534-4fb3-8b76-1e36bd1897bd-kube-api-access-rn99b\") pod \"ovnkube-node-69n7v\" (UID: \"1076c25c-e534-4fb3-8b76-1e36bd1897bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.950217 4857 scope.go:117] "RemoveContainer" containerID="825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.972349 4857 scope.go:117] "RemoveContainer" containerID="1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.991460 4857 scope.go:117] "RemoveContainer" containerID="17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe" Dec 01 21:46:47 crc kubenswrapper[4857]: I1201 21:46:47.997972 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.007914 4857 scope.go:117] "RemoveContainer" containerID="8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.028718 4857 scope.go:117] "RemoveContainer" containerID="dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.053412 4857 scope.go:117] "RemoveContainer" containerID="fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.084862 4857 scope.go:117] "RemoveContainer" containerID="1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca" Dec 01 21:46:48 crc kubenswrapper[4857]: E1201 21:46:48.085534 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca\": container with ID starting with 1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca not found: ID does not exist" containerID="1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.085622 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca"} err="failed to get container status \"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca\": rpc error: code = NotFound desc = could not find container \"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca\": container with ID starting with 1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.085678 4857 scope.go:117] "RemoveContainer" containerID="9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411" Dec 01 21:46:48 crc kubenswrapper[4857]: E1201 21:46:48.086483 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\": container with ID starting with 9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411 not found: ID does not exist" containerID="9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.086553 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411"} err="failed to get container status \"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\": rpc error: code = NotFound desc = could not find container \"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\": container with ID starting with 9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.086593 4857 scope.go:117] "RemoveContainer" containerID="49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c" Dec 01 21:46:48 crc kubenswrapper[4857]: E1201 21:46:48.087180 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\": container with ID starting with 49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c not found: ID does not exist" containerID="49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.087279 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c"} err="failed to get container status \"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\": rpc error: code = NotFound desc = could not find container \"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\": container with ID starting with 49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.087347 4857 scope.go:117] "RemoveContainer" containerID="49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3" Dec 01 21:46:48 crc kubenswrapper[4857]: E1201 21:46:48.087846 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\": container with ID starting with 49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3 not found: ID does not exist" containerID="49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.087906 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3"} err="failed to get container status \"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\": rpc error: code = NotFound desc = could not find container \"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\": container with ID starting with 49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.087942 4857 scope.go:117] "RemoveContainer" containerID="825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462" Dec 01 21:46:48 crc kubenswrapper[4857]: E1201 21:46:48.088391 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\": container with ID starting with 825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462 not found: ID does not exist" containerID="825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.088444 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462"} err="failed to get container status \"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\": rpc error: code = NotFound desc = could not find container \"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\": container with ID starting with 825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.088479 4857 scope.go:117] "RemoveContainer" containerID="1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723" Dec 01 21:46:48 crc kubenswrapper[4857]: E1201 21:46:48.088902 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\": container with ID starting with 1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723 not found: ID does not exist" containerID="1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.088964 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723"} err="failed to get container status \"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\": rpc error: code = NotFound desc = could not find container \"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\": container with ID starting with 1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.089013 4857 scope.go:117] "RemoveContainer" containerID="17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe" Dec 01 21:46:48 crc kubenswrapper[4857]: E1201 21:46:48.089472 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\": container with ID starting with 17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe not found: ID does not exist" containerID="17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.089510 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe"} err="failed to get container status \"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\": rpc error: code = NotFound desc = could not find container \"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\": container with ID starting with 17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.089540 4857 scope.go:117] "RemoveContainer" containerID="8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078" Dec 01 21:46:48 crc kubenswrapper[4857]: E1201 21:46:48.089949 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\": container with ID starting with 8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078 not found: ID does not exist" containerID="8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.090011 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078"} err="failed to get container status \"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\": rpc error: code = NotFound desc = could not find container \"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\": container with ID starting with 8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.090088 4857 scope.go:117] "RemoveContainer" containerID="dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d" Dec 01 21:46:48 crc kubenswrapper[4857]: E1201 21:46:48.090775 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\": container with ID starting with dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d not found: ID does not exist" containerID="dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.090840 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d"} err="failed to get container status \"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\": rpc error: code = NotFound desc = could not find container \"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\": container with ID starting with dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.090973 4857 scope.go:117] "RemoveContainer" containerID="fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f" Dec 01 21:46:48 crc kubenswrapper[4857]: E1201 21:46:48.091580 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\": container with ID starting with fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f not found: ID does not exist" containerID="fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.091643 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f"} err="failed to get container status \"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\": rpc error: code = NotFound desc = could not find container \"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\": container with ID starting with fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.091681 4857 scope.go:117] "RemoveContainer" containerID="1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.092190 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca"} err="failed to get container status \"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca\": rpc error: code = NotFound desc = could not find container \"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca\": container with ID starting with 1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.092258 4857 scope.go:117] "RemoveContainer" containerID="9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.092698 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411"} err="failed to get container status \"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\": rpc error: code = NotFound desc = could not find container \"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\": container with ID starting with 9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.092755 4857 scope.go:117] "RemoveContainer" containerID="49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.093212 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c"} err="failed to get container status \"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\": rpc error: code = NotFound desc = could not find container \"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\": container with ID starting with 49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.093267 4857 scope.go:117] "RemoveContainer" containerID="49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.093627 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3"} err="failed to get container status \"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\": rpc error: code = NotFound desc = could not find container \"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\": container with ID starting with 49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.093716 4857 scope.go:117] "RemoveContainer" containerID="825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.094125 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462"} err="failed to get container status \"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\": rpc error: code = NotFound desc = could not find container \"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\": container with ID starting with 825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.094179 4857 scope.go:117] "RemoveContainer" containerID="1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.094504 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723"} err="failed to get container status \"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\": rpc error: code = NotFound desc = could not find container \"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\": container with ID starting with 1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.094530 4857 scope.go:117] "RemoveContainer" containerID="17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.094847 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe"} err="failed to get container status \"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\": rpc error: code = NotFound desc = could not find container \"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\": container with ID starting with 17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.094890 4857 scope.go:117] "RemoveContainer" containerID="8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.095247 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078"} err="failed to get container status \"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\": rpc error: code = NotFound desc = could not find container \"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\": container with ID starting with 8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.095286 4857 scope.go:117] "RemoveContainer" containerID="dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.095606 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d"} err="failed to get container status \"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\": rpc error: code = NotFound desc = could not find container \"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\": container with ID starting with dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.095644 4857 scope.go:117] "RemoveContainer" containerID="fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.095956 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f"} err="failed to get container status \"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\": rpc error: code = NotFound desc = could not find container \"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\": container with ID starting with fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.096008 4857 scope.go:117] "RemoveContainer" containerID="1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.096385 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca"} err="failed to get container status \"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca\": rpc error: code = NotFound desc = could not find container \"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca\": container with ID starting with 1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.096420 4857 scope.go:117] "RemoveContainer" containerID="9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.096714 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411"} err="failed to get container status \"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\": rpc error: code = NotFound desc = could not find container \"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\": container with ID starting with 9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.096760 4857 scope.go:117] "RemoveContainer" containerID="49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.097142 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c"} err="failed to get container status \"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\": rpc error: code = NotFound desc = could not find container \"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\": container with ID starting with 49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.097177 4857 scope.go:117] "RemoveContainer" containerID="49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.097838 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3"} err="failed to get container status \"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\": rpc error: code = NotFound desc = could not find container \"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\": container with ID starting with 49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.097865 4857 scope.go:117] "RemoveContainer" containerID="825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.098224 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462"} err="failed to get container status \"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\": rpc error: code = NotFound desc = could not find container \"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\": container with ID starting with 825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.098251 4857 scope.go:117] "RemoveContainer" containerID="1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.098597 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723"} err="failed to get container status \"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\": rpc error: code = NotFound desc = could not find container \"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\": container with ID starting with 1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.098616 4857 scope.go:117] "RemoveContainer" containerID="17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.098877 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe"} err="failed to get container status \"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\": rpc error: code = NotFound desc = could not find container \"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\": container with ID starting with 17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.098893 4857 scope.go:117] "RemoveContainer" containerID="8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.099192 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078"} err="failed to get container status \"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\": rpc error: code = NotFound desc = could not find container \"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\": container with ID starting with 8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.099211 4857 scope.go:117] "RemoveContainer" containerID="dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.099435 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d"} err="failed to get container status \"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\": rpc error: code = NotFound desc = could not find container \"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\": container with ID starting with dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.099457 4857 scope.go:117] "RemoveContainer" containerID="fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.099733 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f"} err="failed to get container status \"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\": rpc error: code = NotFound desc = could not find container \"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\": container with ID starting with fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.099750 4857 scope.go:117] "RemoveContainer" containerID="1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.100113 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca"} err="failed to get container status \"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca\": rpc error: code = NotFound desc = could not find container \"1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca\": container with ID starting with 1c2b6781be30d6719a6253ed7871ff7d098a88a3eafb4df63642e042e0bac7ca not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.100167 4857 scope.go:117] "RemoveContainer" containerID="9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.100673 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411"} err="failed to get container status \"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\": rpc error: code = NotFound desc = could not find container \"9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411\": container with ID starting with 9f5f6f5f7548a1b6a6c1360352e895e5ec8605129fc8063107a0b985b4a0f411 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.100725 4857 scope.go:117] "RemoveContainer" containerID="49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.101120 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c"} err="failed to get container status \"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\": rpc error: code = NotFound desc = could not find container \"49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c\": container with ID starting with 49f8e850f3cb24db4cd087de9d8d61618a92d9540430760984e09bbecc39f66c not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.101151 4857 scope.go:117] "RemoveContainer" containerID="49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.101451 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3"} err="failed to get container status \"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\": rpc error: code = NotFound desc = could not find container \"49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3\": container with ID starting with 49335d8727dd01432d3aeeb2525f78bb6cd01589fe7127b277dfb15b43e239e3 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.101479 4857 scope.go:117] "RemoveContainer" containerID="825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.101777 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462"} err="failed to get container status \"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\": rpc error: code = NotFound desc = could not find container \"825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462\": container with ID starting with 825c1096c4867727d3c4e072a8846bf0c61b8b568c68627937eec69f41304462 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.101803 4857 scope.go:117] "RemoveContainer" containerID="1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.102091 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723"} err="failed to get container status \"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\": rpc error: code = NotFound desc = could not find container \"1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723\": container with ID starting with 1fb65ac4f6640d0463ebb9cb1c2bda274abb606f07073ce82519f740b0c16723 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.102117 4857 scope.go:117] "RemoveContainer" containerID="17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.102499 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe"} err="failed to get container status \"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\": rpc error: code = NotFound desc = could not find container \"17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe\": container with ID starting with 17039afe806d2692848732037424fc83de3b5416a4998f89e02f575ae98bd6fe not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.102544 4857 scope.go:117] "RemoveContainer" containerID="8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.102997 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078"} err="failed to get container status \"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\": rpc error: code = NotFound desc = could not find container \"8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078\": container with ID starting with 8c86748c057c74dcc4b324ed9d19c6a4bd151c2d2f3b780933a707a47a8a4078 not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.103030 4857 scope.go:117] "RemoveContainer" containerID="dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.103400 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d"} err="failed to get container status \"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\": rpc error: code = NotFound desc = could not find container \"dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d\": container with ID starting with dc3e5b0d2c7398f2b42f000b6d8631e57b7c0dc56c7df5b38577f003dbb62c9d not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.103428 4857 scope.go:117] "RemoveContainer" containerID="fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.103754 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f"} err="failed to get container status \"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\": rpc error: code = NotFound desc = could not find container \"fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f\": container with ID starting with fe7ab4e2bb4bf905f1d09362af35fa4a15b1ff6e0fd27e7c903e33297c50941f not found: ID does not exist" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.861175 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dfxtl_fc7d55e5-8085-40b7-8147-8508ee6c3f61/kube-multus/2.log" Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.865486 4857 generic.go:334] "Generic (PLEG): container finished" podID="1076c25c-e534-4fb3-8b76-1e36bd1897bd" containerID="277db58b6e1ce3c134130cfbbdb31cba6732119a73a69e6eb33c87d4d381b53e" exitCode=0 Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.865560 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" event={"ID":"1076c25c-e534-4fb3-8b76-1e36bd1897bd","Type":"ContainerDied","Data":"277db58b6e1ce3c134130cfbbdb31cba6732119a73a69e6eb33c87d4d381b53e"} Dec 01 21:46:48 crc kubenswrapper[4857]: I1201 21:46:48.865623 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" event={"ID":"1076c25c-e534-4fb3-8b76-1e36bd1897bd","Type":"ContainerStarted","Data":"313e82be25e571f12e1ac11f57e8c3064736ba8f95ff7063593c761c5163a4e7"} Dec 01 21:46:49 crc kubenswrapper[4857]: I1201 21:46:49.851655 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="557fd3cd-9ab9-4135-af14-f748fdc34a90" path="/var/lib/kubelet/pods/557fd3cd-9ab9-4135-af14-f748fdc34a90/volumes" Dec 01 21:46:49 crc kubenswrapper[4857]: I1201 21:46:49.880521 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" event={"ID":"1076c25c-e534-4fb3-8b76-1e36bd1897bd","Type":"ContainerStarted","Data":"7004d6d3cbaf3db05c9b80fa9724c06c85fd3308edb2d5de6187e7082e665264"} Dec 01 21:46:49 crc kubenswrapper[4857]: I1201 21:46:49.880704 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" event={"ID":"1076c25c-e534-4fb3-8b76-1e36bd1897bd","Type":"ContainerStarted","Data":"6e3bf66615a07befe38bd0de15874849412303b3a8f3900c215f81b6539cee69"} Dec 01 21:46:49 crc kubenswrapper[4857]: I1201 21:46:49.880844 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" event={"ID":"1076c25c-e534-4fb3-8b76-1e36bd1897bd","Type":"ContainerStarted","Data":"99b1a7fb17823877fa49159647d290d3bc97bc8ea3c040556072d67cce6379be"} Dec 01 21:46:49 crc kubenswrapper[4857]: I1201 21:46:49.880970 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" event={"ID":"1076c25c-e534-4fb3-8b76-1e36bd1897bd","Type":"ContainerStarted","Data":"fe8b687601de6bfe97de6285a0cd31c34fe036630648d2978d5d2cb05832020d"} Dec 01 21:46:49 crc kubenswrapper[4857]: I1201 21:46:49.881161 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" event={"ID":"1076c25c-e534-4fb3-8b76-1e36bd1897bd","Type":"ContainerStarted","Data":"22933503fbd149c6975a9acd71166ba9a39dcc756b9698c17336e994410496b4"} Dec 01 21:46:50 crc kubenswrapper[4857]: I1201 21:46:50.905822 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" event={"ID":"1076c25c-e534-4fb3-8b76-1e36bd1897bd","Type":"ContainerStarted","Data":"dea9b5af1199de05c3fc63c53f2a62136db97791063c0c20724da68c8f0d2af7"} Dec 01 21:46:52 crc kubenswrapper[4857]: I1201 21:46:52.920299 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" event={"ID":"1076c25c-e534-4fb3-8b76-1e36bd1897bd","Type":"ContainerStarted","Data":"b48a9d1c13adc489376ec91ed9a04e7153a116eff05b7d65f934d7e78aab677a"} Dec 01 21:46:54 crc kubenswrapper[4857]: I1201 21:46:54.949282 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" event={"ID":"1076c25c-e534-4fb3-8b76-1e36bd1897bd","Type":"ContainerStarted","Data":"f67768ee5712564cd87fd220e896b8570dc478c71d09f9f455fda1040337c2a9"} Dec 01 21:46:54 crc kubenswrapper[4857]: I1201 21:46:54.950020 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:54 crc kubenswrapper[4857]: I1201 21:46:54.950171 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:54 crc kubenswrapper[4857]: I1201 21:46:54.950414 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:54 crc kubenswrapper[4857]: I1201 21:46:54.981903 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" podStartSLOduration=7.981885832 podStartE2EDuration="7.981885832s" podCreationTimestamp="2025-12-01 21:46:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:46:54.979075772 +0000 UTC m=+733.469138089" watchObservedRunningTime="2025-12-01 21:46:54.981885832 +0000 UTC m=+733.471948159" Dec 01 21:46:54 crc kubenswrapper[4857]: I1201 21:46:54.985179 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:46:54 crc kubenswrapper[4857]: I1201 21:46:54.988076 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:47:00 crc kubenswrapper[4857]: I1201 21:47:00.836315 4857 scope.go:117] "RemoveContainer" containerID="0d75201bf392092adaf26780aa1582bd7f68dae7e1f0c99d67785a27dd80b8f8" Dec 01 21:47:02 crc kubenswrapper[4857]: I1201 21:47:02.005455 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dfxtl_fc7d55e5-8085-40b7-8147-8508ee6c3f61/kube-multus/2.log" Dec 01 21:47:02 crc kubenswrapper[4857]: I1201 21:47:02.006013 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dfxtl" event={"ID":"fc7d55e5-8085-40b7-8147-8508ee6c3f61","Type":"ContainerStarted","Data":"534642a33ad83547d7ad85d4494ffbce00da6a5d683aedad3e8543b85b49a775"} Dec 01 21:47:18 crc kubenswrapper[4857]: I1201 21:47:18.038456 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-69n7v" Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.573478 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk"] Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.576200 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.587197 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk"] Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.590076 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.590790 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q27dp\" (UniqueName: \"kubernetes.io/projected/d0bf962d-59bc-40cf-b395-e5cd376eda23-kube-api-access-q27dp\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk\" (UID: \"d0bf962d-59bc-40cf-b395-e5cd376eda23\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.590882 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0bf962d-59bc-40cf-b395-e5cd376eda23-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk\" (UID: \"d0bf962d-59bc-40cf-b395-e5cd376eda23\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.590935 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0bf962d-59bc-40cf-b395-e5cd376eda23-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk\" (UID: \"d0bf962d-59bc-40cf-b395-e5cd376eda23\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.691943 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0bf962d-59bc-40cf-b395-e5cd376eda23-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk\" (UID: \"d0bf962d-59bc-40cf-b395-e5cd376eda23\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.692059 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q27dp\" (UniqueName: \"kubernetes.io/projected/d0bf962d-59bc-40cf-b395-e5cd376eda23-kube-api-access-q27dp\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk\" (UID: \"d0bf962d-59bc-40cf-b395-e5cd376eda23\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.692100 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0bf962d-59bc-40cf-b395-e5cd376eda23-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk\" (UID: \"d0bf962d-59bc-40cf-b395-e5cd376eda23\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.694368 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0bf962d-59bc-40cf-b395-e5cd376eda23-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk\" (UID: \"d0bf962d-59bc-40cf-b395-e5cd376eda23\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.694548 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0bf962d-59bc-40cf-b395-e5cd376eda23-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk\" (UID: \"d0bf962d-59bc-40cf-b395-e5cd376eda23\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.711581 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q27dp\" (UniqueName: \"kubernetes.io/projected/d0bf962d-59bc-40cf-b395-e5cd376eda23-kube-api-access-q27dp\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk\" (UID: \"d0bf962d-59bc-40cf-b395-e5cd376eda23\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:25 crc kubenswrapper[4857]: I1201 21:47:25.936815 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:26 crc kubenswrapper[4857]: I1201 21:47:26.220755 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk"] Dec 01 21:47:26 crc kubenswrapper[4857]: I1201 21:47:26.738781 4857 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 21:47:27 crc kubenswrapper[4857]: I1201 21:47:27.190943 4857 generic.go:334] "Generic (PLEG): container finished" podID="d0bf962d-59bc-40cf-b395-e5cd376eda23" containerID="d52aa8a011d75358fe1aaf0fe398725bb16bc2acde6b7f037d224ead222b8a0f" exitCode=0 Dec 01 21:47:27 crc kubenswrapper[4857]: I1201 21:47:27.190988 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" event={"ID":"d0bf962d-59bc-40cf-b395-e5cd376eda23","Type":"ContainerDied","Data":"d52aa8a011d75358fe1aaf0fe398725bb16bc2acde6b7f037d224ead222b8a0f"} Dec 01 21:47:27 crc kubenswrapper[4857]: I1201 21:47:27.191033 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" event={"ID":"d0bf962d-59bc-40cf-b395-e5cd376eda23","Type":"ContainerStarted","Data":"a5431ff0dac87ba5a5f6efead921197c786b7bf5b9332d0dc4103da6ea32b2f8"} Dec 01 21:47:27 crc kubenswrapper[4857]: I1201 21:47:27.906184 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wgwtn"] Dec 01 21:47:27 crc kubenswrapper[4857]: I1201 21:47:27.907944 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:27 crc kubenswrapper[4857]: I1201 21:47:27.924967 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e654be-8e0b-4421-8d29-c9708ba8af3e-catalog-content\") pod \"redhat-operators-wgwtn\" (UID: \"49e654be-8e0b-4421-8d29-c9708ba8af3e\") " pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:27 crc kubenswrapper[4857]: I1201 21:47:27.925155 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e654be-8e0b-4421-8d29-c9708ba8af3e-utilities\") pod \"redhat-operators-wgwtn\" (UID: \"49e654be-8e0b-4421-8d29-c9708ba8af3e\") " pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:27 crc kubenswrapper[4857]: I1201 21:47:27.925232 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2jx7\" (UniqueName: \"kubernetes.io/projected/49e654be-8e0b-4421-8d29-c9708ba8af3e-kube-api-access-p2jx7\") pod \"redhat-operators-wgwtn\" (UID: \"49e654be-8e0b-4421-8d29-c9708ba8af3e\") " pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:27 crc kubenswrapper[4857]: I1201 21:47:27.931679 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wgwtn"] Dec 01 21:47:28 crc kubenswrapper[4857]: I1201 21:47:28.027139 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e654be-8e0b-4421-8d29-c9708ba8af3e-utilities\") pod \"redhat-operators-wgwtn\" (UID: \"49e654be-8e0b-4421-8d29-c9708ba8af3e\") " pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:28 crc kubenswrapper[4857]: I1201 21:47:28.029230 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2jx7\" (UniqueName: \"kubernetes.io/projected/49e654be-8e0b-4421-8d29-c9708ba8af3e-kube-api-access-p2jx7\") pod \"redhat-operators-wgwtn\" (UID: \"49e654be-8e0b-4421-8d29-c9708ba8af3e\") " pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:28 crc kubenswrapper[4857]: I1201 21:47:28.029335 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e654be-8e0b-4421-8d29-c9708ba8af3e-catalog-content\") pod \"redhat-operators-wgwtn\" (UID: \"49e654be-8e0b-4421-8d29-c9708ba8af3e\") " pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:28 crc kubenswrapper[4857]: I1201 21:47:28.030079 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e654be-8e0b-4421-8d29-c9708ba8af3e-utilities\") pod \"redhat-operators-wgwtn\" (UID: \"49e654be-8e0b-4421-8d29-c9708ba8af3e\") " pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:28 crc kubenswrapper[4857]: I1201 21:47:28.030582 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e654be-8e0b-4421-8d29-c9708ba8af3e-catalog-content\") pod \"redhat-operators-wgwtn\" (UID: \"49e654be-8e0b-4421-8d29-c9708ba8af3e\") " pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:28 crc kubenswrapper[4857]: I1201 21:47:28.061374 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2jx7\" (UniqueName: \"kubernetes.io/projected/49e654be-8e0b-4421-8d29-c9708ba8af3e-kube-api-access-p2jx7\") pod \"redhat-operators-wgwtn\" (UID: \"49e654be-8e0b-4421-8d29-c9708ba8af3e\") " pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:28 crc kubenswrapper[4857]: I1201 21:47:28.231347 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:28 crc kubenswrapper[4857]: I1201 21:47:28.428256 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wgwtn"] Dec 01 21:47:28 crc kubenswrapper[4857]: W1201 21:47:28.480131 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49e654be_8e0b_4421_8d29_c9708ba8af3e.slice/crio-831500b90a942cf86e158d890608e6e62762cc7c8fe27f1a80ae01edfacae391 WatchSource:0}: Error finding container 831500b90a942cf86e158d890608e6e62762cc7c8fe27f1a80ae01edfacae391: Status 404 returned error can't find the container with id 831500b90a942cf86e158d890608e6e62762cc7c8fe27f1a80ae01edfacae391 Dec 01 21:47:29 crc kubenswrapper[4857]: I1201 21:47:29.204389 4857 generic.go:334] "Generic (PLEG): container finished" podID="d0bf962d-59bc-40cf-b395-e5cd376eda23" containerID="40781889806aa5c1fc0d884aeecf4891ce872af827d43007a65b093208f47cae" exitCode=0 Dec 01 21:47:29 crc kubenswrapper[4857]: I1201 21:47:29.204452 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" event={"ID":"d0bf962d-59bc-40cf-b395-e5cd376eda23","Type":"ContainerDied","Data":"40781889806aa5c1fc0d884aeecf4891ce872af827d43007a65b093208f47cae"} Dec 01 21:47:29 crc kubenswrapper[4857]: I1201 21:47:29.206751 4857 generic.go:334] "Generic (PLEG): container finished" podID="49e654be-8e0b-4421-8d29-c9708ba8af3e" containerID="2941770c7f5f647f417d319bb3728ce80d5f8612b48ed35941b8bf80fd065356" exitCode=0 Dec 01 21:47:29 crc kubenswrapper[4857]: I1201 21:47:29.206775 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgwtn" event={"ID":"49e654be-8e0b-4421-8d29-c9708ba8af3e","Type":"ContainerDied","Data":"2941770c7f5f647f417d319bb3728ce80d5f8612b48ed35941b8bf80fd065356"} Dec 01 21:47:29 crc kubenswrapper[4857]: I1201 21:47:29.206793 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgwtn" event={"ID":"49e654be-8e0b-4421-8d29-c9708ba8af3e","Type":"ContainerStarted","Data":"831500b90a942cf86e158d890608e6e62762cc7c8fe27f1a80ae01edfacae391"} Dec 01 21:47:30 crc kubenswrapper[4857]: I1201 21:47:30.217930 4857 generic.go:334] "Generic (PLEG): container finished" podID="d0bf962d-59bc-40cf-b395-e5cd376eda23" containerID="cd625f17fb6eeb33226a7bdc2a5cc59043e4969d1d2c19f1c03f6ec251e6b7e4" exitCode=0 Dec 01 21:47:30 crc kubenswrapper[4857]: I1201 21:47:30.219540 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" event={"ID":"d0bf962d-59bc-40cf-b395-e5cd376eda23","Type":"ContainerDied","Data":"cd625f17fb6eeb33226a7bdc2a5cc59043e4969d1d2c19f1c03f6ec251e6b7e4"} Dec 01 21:47:30 crc kubenswrapper[4857]: I1201 21:47:30.223131 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgwtn" event={"ID":"49e654be-8e0b-4421-8d29-c9708ba8af3e","Type":"ContainerStarted","Data":"bda606175e68af92d9c75538f4bb325cf5e3e6792e9751ba83f4ce9be5dd2887"} Dec 01 21:47:31 crc kubenswrapper[4857]: I1201 21:47:31.551425 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:31 crc kubenswrapper[4857]: I1201 21:47:31.682983 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0bf962d-59bc-40cf-b395-e5cd376eda23-bundle\") pod \"d0bf962d-59bc-40cf-b395-e5cd376eda23\" (UID: \"d0bf962d-59bc-40cf-b395-e5cd376eda23\") " Dec 01 21:47:31 crc kubenswrapper[4857]: I1201 21:47:31.683076 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0bf962d-59bc-40cf-b395-e5cd376eda23-util\") pod \"d0bf962d-59bc-40cf-b395-e5cd376eda23\" (UID: \"d0bf962d-59bc-40cf-b395-e5cd376eda23\") " Dec 01 21:47:31 crc kubenswrapper[4857]: I1201 21:47:31.683115 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q27dp\" (UniqueName: \"kubernetes.io/projected/d0bf962d-59bc-40cf-b395-e5cd376eda23-kube-api-access-q27dp\") pod \"d0bf962d-59bc-40cf-b395-e5cd376eda23\" (UID: \"d0bf962d-59bc-40cf-b395-e5cd376eda23\") " Dec 01 21:47:31 crc kubenswrapper[4857]: I1201 21:47:31.684023 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0bf962d-59bc-40cf-b395-e5cd376eda23-bundle" (OuterVolumeSpecName: "bundle") pod "d0bf962d-59bc-40cf-b395-e5cd376eda23" (UID: "d0bf962d-59bc-40cf-b395-e5cd376eda23"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:47:31 crc kubenswrapper[4857]: I1201 21:47:31.691897 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0bf962d-59bc-40cf-b395-e5cd376eda23-kube-api-access-q27dp" (OuterVolumeSpecName: "kube-api-access-q27dp") pod "d0bf962d-59bc-40cf-b395-e5cd376eda23" (UID: "d0bf962d-59bc-40cf-b395-e5cd376eda23"). InnerVolumeSpecName "kube-api-access-q27dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:47:31 crc kubenswrapper[4857]: I1201 21:47:31.784787 4857 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0bf962d-59bc-40cf-b395-e5cd376eda23-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:47:31 crc kubenswrapper[4857]: I1201 21:47:31.784830 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q27dp\" (UniqueName: \"kubernetes.io/projected/d0bf962d-59bc-40cf-b395-e5cd376eda23-kube-api-access-q27dp\") on node \"crc\" DevicePath \"\"" Dec 01 21:47:31 crc kubenswrapper[4857]: I1201 21:47:31.789003 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0bf962d-59bc-40cf-b395-e5cd376eda23-util" (OuterVolumeSpecName: "util") pod "d0bf962d-59bc-40cf-b395-e5cd376eda23" (UID: "d0bf962d-59bc-40cf-b395-e5cd376eda23"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:47:31 crc kubenswrapper[4857]: I1201 21:47:31.885917 4857 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0bf962d-59bc-40cf-b395-e5cd376eda23-util\") on node \"crc\" DevicePath \"\"" Dec 01 21:47:32 crc kubenswrapper[4857]: I1201 21:47:32.237362 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" event={"ID":"d0bf962d-59bc-40cf-b395-e5cd376eda23","Type":"ContainerDied","Data":"a5431ff0dac87ba5a5f6efead921197c786b7bf5b9332d0dc4103da6ea32b2f8"} Dec 01 21:47:32 crc kubenswrapper[4857]: I1201 21:47:32.237419 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5431ff0dac87ba5a5f6efead921197c786b7bf5b9332d0dc4103da6ea32b2f8" Dec 01 21:47:32 crc kubenswrapper[4857]: I1201 21:47:32.237512 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk" Dec 01 21:47:32 crc kubenswrapper[4857]: I1201 21:47:32.242503 4857 generic.go:334] "Generic (PLEG): container finished" podID="49e654be-8e0b-4421-8d29-c9708ba8af3e" containerID="bda606175e68af92d9c75538f4bb325cf5e3e6792e9751ba83f4ce9be5dd2887" exitCode=0 Dec 01 21:47:32 crc kubenswrapper[4857]: I1201 21:47:32.242558 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgwtn" event={"ID":"49e654be-8e0b-4421-8d29-c9708ba8af3e","Type":"ContainerDied","Data":"bda606175e68af92d9c75538f4bb325cf5e3e6792e9751ba83f4ce9be5dd2887"} Dec 01 21:47:33 crc kubenswrapper[4857]: I1201 21:47:33.251123 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgwtn" event={"ID":"49e654be-8e0b-4421-8d29-c9708ba8af3e","Type":"ContainerStarted","Data":"e73de1d80b81f56a349fb898eac2fb806d6c36c416ab2aa788517d3eadadf750"} Dec 01 21:47:33 crc kubenswrapper[4857]: I1201 21:47:33.281612 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wgwtn" podStartSLOduration=2.853397794 podStartE2EDuration="6.281583026s" podCreationTimestamp="2025-12-01 21:47:27 +0000 UTC" firstStartedPulling="2025-12-01 21:47:29.207815072 +0000 UTC m=+767.697877389" lastFinishedPulling="2025-12-01 21:47:32.636000294 +0000 UTC m=+771.126062621" observedRunningTime="2025-12-01 21:47:33.277771731 +0000 UTC m=+771.767834088" watchObservedRunningTime="2025-12-01 21:47:33.281583026 +0000 UTC m=+771.771645363" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.522937 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-m8q4r"] Dec 01 21:47:35 crc kubenswrapper[4857]: E1201 21:47:35.523598 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0bf962d-59bc-40cf-b395-e5cd376eda23" containerName="extract" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.523614 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0bf962d-59bc-40cf-b395-e5cd376eda23" containerName="extract" Dec 01 21:47:35 crc kubenswrapper[4857]: E1201 21:47:35.523643 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0bf962d-59bc-40cf-b395-e5cd376eda23" containerName="util" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.523651 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0bf962d-59bc-40cf-b395-e5cd376eda23" containerName="util" Dec 01 21:47:35 crc kubenswrapper[4857]: E1201 21:47:35.523664 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0bf962d-59bc-40cf-b395-e5cd376eda23" containerName="pull" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.523670 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0bf962d-59bc-40cf-b395-e5cd376eda23" containerName="pull" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.523778 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0bf962d-59bc-40cf-b395-e5cd376eda23" containerName="extract" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.524264 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-m8q4r" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.532166 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.532376 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-hmd9m" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.532499 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.538236 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-m8q4r"] Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.656943 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdkv5\" (UniqueName: \"kubernetes.io/projected/8c989c5f-b814-437a-bdca-86e63044ce91-kube-api-access-tdkv5\") pod \"nmstate-operator-5b5b58f5c8-m8q4r\" (UID: \"8c989c5f-b814-437a-bdca-86e63044ce91\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-m8q4r" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.757930 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdkv5\" (UniqueName: \"kubernetes.io/projected/8c989c5f-b814-437a-bdca-86e63044ce91-kube-api-access-tdkv5\") pod \"nmstate-operator-5b5b58f5c8-m8q4r\" (UID: \"8c989c5f-b814-437a-bdca-86e63044ce91\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-m8q4r" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.785202 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdkv5\" (UniqueName: \"kubernetes.io/projected/8c989c5f-b814-437a-bdca-86e63044ce91-kube-api-access-tdkv5\") pod \"nmstate-operator-5b5b58f5c8-m8q4r\" (UID: \"8c989c5f-b814-437a-bdca-86e63044ce91\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-m8q4r" Dec 01 21:47:35 crc kubenswrapper[4857]: I1201 21:47:35.842842 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-m8q4r" Dec 01 21:47:36 crc kubenswrapper[4857]: I1201 21:47:36.194247 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-m8q4r"] Dec 01 21:47:36 crc kubenswrapper[4857]: I1201 21:47:36.267974 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-m8q4r" event={"ID":"8c989c5f-b814-437a-bdca-86e63044ce91","Type":"ContainerStarted","Data":"7883602ba740b6f57c95464bad2ed83b727494fa68151f6d5cf6223513c1b3fe"} Dec 01 21:47:38 crc kubenswrapper[4857]: I1201 21:47:38.232077 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:38 crc kubenswrapper[4857]: I1201 21:47:38.232682 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:39 crc kubenswrapper[4857]: I1201 21:47:39.275818 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wgwtn" podUID="49e654be-8e0b-4421-8d29-c9708ba8af3e" containerName="registry-server" probeResult="failure" output=< Dec 01 21:47:39 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Dec 01 21:47:39 crc kubenswrapper[4857]: > Dec 01 21:47:41 crc kubenswrapper[4857]: I1201 21:47:41.310931 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-m8q4r" event={"ID":"8c989c5f-b814-437a-bdca-86e63044ce91","Type":"ContainerStarted","Data":"05d948c2b1de9ca5cf163cc31ae4f6f3789a9971c3282926cea0643711b8dd59"} Dec 01 21:47:41 crc kubenswrapper[4857]: I1201 21:47:41.338587 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-m8q4r" podStartSLOduration=2.44938454 podStartE2EDuration="6.338559259s" podCreationTimestamp="2025-12-01 21:47:35 +0000 UTC" firstStartedPulling="2025-12-01 21:47:36.201175647 +0000 UTC m=+774.691237964" lastFinishedPulling="2025-12-01 21:47:40.090350356 +0000 UTC m=+778.580412683" observedRunningTime="2025-12-01 21:47:41.337357289 +0000 UTC m=+779.827419626" watchObservedRunningTime="2025-12-01 21:47:41.338559259 +0000 UTC m=+779.828621616" Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.825642 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w"] Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.827262 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.829561 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-m49jm"] Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.829846 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.830219 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-z2wwl" Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.830600 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m49jm" Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.850426 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w"] Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.854596 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-m49jm"] Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.858082 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-grx52"] Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.858835 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.957789 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8ztf\" (UniqueName: \"kubernetes.io/projected/79431306-c7c5-4c07-8b52-ec94fe45e6da-kube-api-access-z8ztf\") pod \"nmstate-webhook-5f6d4c5ccb-bdx6w\" (UID: \"79431306-c7c5-4c07-8b52-ec94fe45e6da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.957876 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ncgt\" (UniqueName: \"kubernetes.io/projected/c7894b0e-91be-4ed2-810b-f9e24df87fae-kube-api-access-5ncgt\") pod \"nmstate-metrics-7f946cbc9-m49jm\" (UID: \"c7894b0e-91be-4ed2-810b-f9e24df87fae\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m49jm" Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.957965 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/79431306-c7c5-4c07-8b52-ec94fe45e6da-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-bdx6w\" (UID: \"79431306-c7c5-4c07-8b52-ec94fe45e6da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.968672 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv"] Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.969649 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.975275 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.980317 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-8rdpx" Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.981241 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv"] Dec 01 21:47:45 crc kubenswrapper[4857]: I1201 21:47:45.981277 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.059033 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e37b046b-176a-4bc3-bcbc-3093fe08b256-dbus-socket\") pod \"nmstate-handler-grx52\" (UID: \"e37b046b-176a-4bc3-bcbc-3093fe08b256\") " pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.059116 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8ztf\" (UniqueName: \"kubernetes.io/projected/79431306-c7c5-4c07-8b52-ec94fe45e6da-kube-api-access-z8ztf\") pod \"nmstate-webhook-5f6d4c5ccb-bdx6w\" (UID: \"79431306-c7c5-4c07-8b52-ec94fe45e6da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.059152 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ncgt\" (UniqueName: \"kubernetes.io/projected/c7894b0e-91be-4ed2-810b-f9e24df87fae-kube-api-access-5ncgt\") pod \"nmstate-metrics-7f946cbc9-m49jm\" (UID: \"c7894b0e-91be-4ed2-810b-f9e24df87fae\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m49jm" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.059174 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e37b046b-176a-4bc3-bcbc-3093fe08b256-ovs-socket\") pod \"nmstate-handler-grx52\" (UID: \"e37b046b-176a-4bc3-bcbc-3093fe08b256\") " pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.059878 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/79431306-c7c5-4c07-8b52-ec94fe45e6da-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-bdx6w\" (UID: \"79431306-c7c5-4c07-8b52-ec94fe45e6da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.059979 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e37b046b-176a-4bc3-bcbc-3093fe08b256-nmstate-lock\") pod \"nmstate-handler-grx52\" (UID: \"e37b046b-176a-4bc3-bcbc-3093fe08b256\") " pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.060016 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s267g\" (UniqueName: \"kubernetes.io/projected/e37b046b-176a-4bc3-bcbc-3093fe08b256-kube-api-access-s267g\") pod \"nmstate-handler-grx52\" (UID: \"e37b046b-176a-4bc3-bcbc-3093fe08b256\") " pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: E1201 21:47:46.060024 4857 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 01 21:47:46 crc kubenswrapper[4857]: E1201 21:47:46.060130 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79431306-c7c5-4c07-8b52-ec94fe45e6da-tls-key-pair podName:79431306-c7c5-4c07-8b52-ec94fe45e6da nodeName:}" failed. No retries permitted until 2025-12-01 21:47:46.560104266 +0000 UTC m=+785.050166583 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/79431306-c7c5-4c07-8b52-ec94fe45e6da-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-bdx6w" (UID: "79431306-c7c5-4c07-8b52-ec94fe45e6da") : secret "openshift-nmstate-webhook" not found Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.088707 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ncgt\" (UniqueName: \"kubernetes.io/projected/c7894b0e-91be-4ed2-810b-f9e24df87fae-kube-api-access-5ncgt\") pod \"nmstate-metrics-7f946cbc9-m49jm\" (UID: \"c7894b0e-91be-4ed2-810b-f9e24df87fae\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m49jm" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.089829 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8ztf\" (UniqueName: \"kubernetes.io/projected/79431306-c7c5-4c07-8b52-ec94fe45e6da-kube-api-access-z8ztf\") pod \"nmstate-webhook-5f6d4c5ccb-bdx6w\" (UID: \"79431306-c7c5-4c07-8b52-ec94fe45e6da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.161780 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0764bcb8-a3f2-46cd-bceb-f08b51012b7d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-78hdv\" (UID: \"0764bcb8-a3f2-46cd-bceb-f08b51012b7d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.161933 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e37b046b-176a-4bc3-bcbc-3093fe08b256-dbus-socket\") pod \"nmstate-handler-grx52\" (UID: \"e37b046b-176a-4bc3-bcbc-3093fe08b256\") " pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.162001 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e37b046b-176a-4bc3-bcbc-3093fe08b256-ovs-socket\") pod \"nmstate-handler-grx52\" (UID: \"e37b046b-176a-4bc3-bcbc-3093fe08b256\") " pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.162074 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0764bcb8-a3f2-46cd-bceb-f08b51012b7d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-78hdv\" (UID: \"0764bcb8-a3f2-46cd-bceb-f08b51012b7d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.162155 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e37b046b-176a-4bc3-bcbc-3093fe08b256-ovs-socket\") pod \"nmstate-handler-grx52\" (UID: \"e37b046b-176a-4bc3-bcbc-3093fe08b256\") " pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.162233 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e37b046b-176a-4bc3-bcbc-3093fe08b256-nmstate-lock\") pod \"nmstate-handler-grx52\" (UID: \"e37b046b-176a-4bc3-bcbc-3093fe08b256\") " pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.162283 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s267g\" (UniqueName: \"kubernetes.io/projected/e37b046b-176a-4bc3-bcbc-3093fe08b256-kube-api-access-s267g\") pod \"nmstate-handler-grx52\" (UID: \"e37b046b-176a-4bc3-bcbc-3093fe08b256\") " pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.162342 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e37b046b-176a-4bc3-bcbc-3093fe08b256-dbus-socket\") pod \"nmstate-handler-grx52\" (UID: \"e37b046b-176a-4bc3-bcbc-3093fe08b256\") " pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.162357 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn54w\" (UniqueName: \"kubernetes.io/projected/0764bcb8-a3f2-46cd-bceb-f08b51012b7d-kube-api-access-hn54w\") pod \"nmstate-console-plugin-7fbb5f6569-78hdv\" (UID: \"0764bcb8-a3f2-46cd-bceb-f08b51012b7d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.162305 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e37b046b-176a-4bc3-bcbc-3093fe08b256-nmstate-lock\") pod \"nmstate-handler-grx52\" (UID: \"e37b046b-176a-4bc3-bcbc-3093fe08b256\") " pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.167852 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-866cc8d79-t6gkb"] Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.168820 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.170669 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m49jm" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.187578 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-866cc8d79-t6gkb"] Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.192736 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s267g\" (UniqueName: \"kubernetes.io/projected/e37b046b-176a-4bc3-bcbc-3093fe08b256-kube-api-access-s267g\") pod \"nmstate-handler-grx52\" (UID: \"e37b046b-176a-4bc3-bcbc-3093fe08b256\") " pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.263824 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-console-config\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.264181 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfqp5\" (UniqueName: \"kubernetes.io/projected/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-kube-api-access-qfqp5\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.264304 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-service-ca\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.264416 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0764bcb8-a3f2-46cd-bceb-f08b51012b7d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-78hdv\" (UID: \"0764bcb8-a3f2-46cd-bceb-f08b51012b7d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.264548 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-console-serving-cert\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.264850 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn54w\" (UniqueName: \"kubernetes.io/projected/0764bcb8-a3f2-46cd-bceb-f08b51012b7d-kube-api-access-hn54w\") pod \"nmstate-console-plugin-7fbb5f6569-78hdv\" (UID: \"0764bcb8-a3f2-46cd-bceb-f08b51012b7d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.264947 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-oauth-serving-cert\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.264976 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-trusted-ca-bundle\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.265026 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-console-oauth-config\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.265074 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0764bcb8-a3f2-46cd-bceb-f08b51012b7d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-78hdv\" (UID: \"0764bcb8-a3f2-46cd-bceb-f08b51012b7d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.266539 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0764bcb8-a3f2-46cd-bceb-f08b51012b7d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-78hdv\" (UID: \"0764bcb8-a3f2-46cd-bceb-f08b51012b7d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.276222 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0764bcb8-a3f2-46cd-bceb-f08b51012b7d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-78hdv\" (UID: \"0764bcb8-a3f2-46cd-bceb-f08b51012b7d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.285338 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn54w\" (UniqueName: \"kubernetes.io/projected/0764bcb8-a3f2-46cd-bceb-f08b51012b7d-kube-api-access-hn54w\") pod \"nmstate-console-plugin-7fbb5f6569-78hdv\" (UID: \"0764bcb8-a3f2-46cd-bceb-f08b51012b7d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.286658 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.367243 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-console-config\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.367308 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfqp5\" (UniqueName: \"kubernetes.io/projected/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-kube-api-access-qfqp5\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.367325 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-service-ca\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.367358 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-console-serving-cert\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.367410 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-oauth-serving-cert\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.367426 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-trusted-ca-bundle\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.367452 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-console-oauth-config\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.368290 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-console-config\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.369136 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-oauth-serving-cert\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.370671 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-trusted-ca-bundle\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.371657 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-service-ca\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.372362 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-console-serving-cert\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.382083 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-console-oauth-config\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.385356 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfqp5\" (UniqueName: \"kubernetes.io/projected/8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9-kube-api-access-qfqp5\") pod \"console-866cc8d79-t6gkb\" (UID: \"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9\") " pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.480081 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.517572 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv"] Dec 01 21:47:46 crc kubenswrapper[4857]: W1201 21:47:46.520420 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0764bcb8_a3f2_46cd_bceb_f08b51012b7d.slice/crio-1171b7835d3a1679f29e7fde4ade323662b47178607c93996323017a1ac18793 WatchSource:0}: Error finding container 1171b7835d3a1679f29e7fde4ade323662b47178607c93996323017a1ac18793: Status 404 returned error can't find the container with id 1171b7835d3a1679f29e7fde4ade323662b47178607c93996323017a1ac18793 Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.529999 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.570483 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/79431306-c7c5-4c07-8b52-ec94fe45e6da-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-bdx6w\" (UID: \"79431306-c7c5-4c07-8b52-ec94fe45e6da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.578029 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/79431306-c7c5-4c07-8b52-ec94fe45e6da-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-bdx6w\" (UID: \"79431306-c7c5-4c07-8b52-ec94fe45e6da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.672238 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-m49jm"] Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.759735 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" Dec 01 21:47:46 crc kubenswrapper[4857]: I1201 21:47:46.836110 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-866cc8d79-t6gkb"] Dec 01 21:47:46 crc kubenswrapper[4857]: W1201 21:47:46.844755 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b0b5522_6aea_4565_98f2_e1c1cf3ee4d9.slice/crio-310e8b36daa254afeda3bc42862a152256046f76d22de1595dd0167406a1b67e WatchSource:0}: Error finding container 310e8b36daa254afeda3bc42862a152256046f76d22de1595dd0167406a1b67e: Status 404 returned error can't find the container with id 310e8b36daa254afeda3bc42862a152256046f76d22de1595dd0167406a1b67e Dec 01 21:47:47 crc kubenswrapper[4857]: I1201 21:47:47.042685 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w"] Dec 01 21:47:47 crc kubenswrapper[4857]: I1201 21:47:47.351464 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" event={"ID":"0764bcb8-a3f2-46cd-bceb-f08b51012b7d","Type":"ContainerStarted","Data":"1171b7835d3a1679f29e7fde4ade323662b47178607c93996323017a1ac18793"} Dec 01 21:47:47 crc kubenswrapper[4857]: I1201 21:47:47.353602 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" event={"ID":"79431306-c7c5-4c07-8b52-ec94fe45e6da","Type":"ContainerStarted","Data":"c20b78c91411b181d372b4cfa85ba7c5e19ed35c7b024b7804850a660dc6d0e6"} Dec 01 21:47:47 crc kubenswrapper[4857]: I1201 21:47:47.356625 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-866cc8d79-t6gkb" event={"ID":"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9","Type":"ContainerStarted","Data":"310e8b36daa254afeda3bc42862a152256046f76d22de1595dd0167406a1b67e"} Dec 01 21:47:47 crc kubenswrapper[4857]: I1201 21:47:47.358685 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-grx52" event={"ID":"e37b046b-176a-4bc3-bcbc-3093fe08b256","Type":"ContainerStarted","Data":"0dd3edb982ef7cb2bef7f3e1b683bc26f14c48e7eb3fce8c519b8d73dad15361"} Dec 01 21:47:47 crc kubenswrapper[4857]: I1201 21:47:47.360603 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m49jm" event={"ID":"c7894b0e-91be-4ed2-810b-f9e24df87fae","Type":"ContainerStarted","Data":"e209052f7b34557a0416e253833dfbbea4072c018d310779acfe40edddb598e7"} Dec 01 21:47:48 crc kubenswrapper[4857]: I1201 21:47:48.294925 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:48 crc kubenswrapper[4857]: I1201 21:47:48.384314 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-866cc8d79-t6gkb" event={"ID":"8b0b5522-6aea-4565-98f2-e1c1cf3ee4d9","Type":"ContainerStarted","Data":"f10b9bf984eebc0dcfa585d89a29ccc86eef26140e903d4580c399b68fa6a817"} Dec 01 21:47:48 crc kubenswrapper[4857]: I1201 21:47:48.388784 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:48 crc kubenswrapper[4857]: I1201 21:47:48.400715 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-866cc8d79-t6gkb" podStartSLOduration=2.400699865 podStartE2EDuration="2.400699865s" podCreationTimestamp="2025-12-01 21:47:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:47:48.400061829 +0000 UTC m=+786.890124146" watchObservedRunningTime="2025-12-01 21:47:48.400699865 +0000 UTC m=+786.890762182" Dec 01 21:47:48 crc kubenswrapper[4857]: I1201 21:47:48.564065 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wgwtn"] Dec 01 21:47:49 crc kubenswrapper[4857]: I1201 21:47:49.401661 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wgwtn" podUID="49e654be-8e0b-4421-8d29-c9708ba8af3e" containerName="registry-server" containerID="cri-o://e73de1d80b81f56a349fb898eac2fb806d6c36c416ab2aa788517d3eadadf750" gracePeriod=2 Dec 01 21:47:50 crc kubenswrapper[4857]: I1201 21:47:50.449361 4857 generic.go:334] "Generic (PLEG): container finished" podID="49e654be-8e0b-4421-8d29-c9708ba8af3e" containerID="e73de1d80b81f56a349fb898eac2fb806d6c36c416ab2aa788517d3eadadf750" exitCode=0 Dec 01 21:47:50 crc kubenswrapper[4857]: I1201 21:47:50.449439 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgwtn" event={"ID":"49e654be-8e0b-4421-8d29-c9708ba8af3e","Type":"ContainerDied","Data":"e73de1d80b81f56a349fb898eac2fb806d6c36c416ab2aa788517d3eadadf750"} Dec 01 21:47:50 crc kubenswrapper[4857]: I1201 21:47:50.813169 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:50 crc kubenswrapper[4857]: I1201 21:47:50.943774 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e654be-8e0b-4421-8d29-c9708ba8af3e-utilities\") pod \"49e654be-8e0b-4421-8d29-c9708ba8af3e\" (UID: \"49e654be-8e0b-4421-8d29-c9708ba8af3e\") " Dec 01 21:47:50 crc kubenswrapper[4857]: I1201 21:47:50.943856 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e654be-8e0b-4421-8d29-c9708ba8af3e-catalog-content\") pod \"49e654be-8e0b-4421-8d29-c9708ba8af3e\" (UID: \"49e654be-8e0b-4421-8d29-c9708ba8af3e\") " Dec 01 21:47:50 crc kubenswrapper[4857]: I1201 21:47:50.943931 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2jx7\" (UniqueName: \"kubernetes.io/projected/49e654be-8e0b-4421-8d29-c9708ba8af3e-kube-api-access-p2jx7\") pod \"49e654be-8e0b-4421-8d29-c9708ba8af3e\" (UID: \"49e654be-8e0b-4421-8d29-c9708ba8af3e\") " Dec 01 21:47:50 crc kubenswrapper[4857]: I1201 21:47:50.944835 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49e654be-8e0b-4421-8d29-c9708ba8af3e-utilities" (OuterVolumeSpecName: "utilities") pod "49e654be-8e0b-4421-8d29-c9708ba8af3e" (UID: "49e654be-8e0b-4421-8d29-c9708ba8af3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:47:50 crc kubenswrapper[4857]: I1201 21:47:50.949832 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e654be-8e0b-4421-8d29-c9708ba8af3e-kube-api-access-p2jx7" (OuterVolumeSpecName: "kube-api-access-p2jx7") pod "49e654be-8e0b-4421-8d29-c9708ba8af3e" (UID: "49e654be-8e0b-4421-8d29-c9708ba8af3e"). InnerVolumeSpecName "kube-api-access-p2jx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.045590 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e654be-8e0b-4421-8d29-c9708ba8af3e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.048599 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2jx7\" (UniqueName: \"kubernetes.io/projected/49e654be-8e0b-4421-8d29-c9708ba8af3e-kube-api-access-p2jx7\") on node \"crc\" DevicePath \"\"" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.052548 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49e654be-8e0b-4421-8d29-c9708ba8af3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49e654be-8e0b-4421-8d29-c9708ba8af3e" (UID: "49e654be-8e0b-4421-8d29-c9708ba8af3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.150164 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e654be-8e0b-4421-8d29-c9708ba8af3e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.465720 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" event={"ID":"79431306-c7c5-4c07-8b52-ec94fe45e6da","Type":"ContainerStarted","Data":"79a8884829eab3df05e60bb15b9d833f1388e23e4859c1bc4e465501d2a9aeef"} Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.467138 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.472164 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgwtn" event={"ID":"49e654be-8e0b-4421-8d29-c9708ba8af3e","Type":"ContainerDied","Data":"831500b90a942cf86e158d890608e6e62762cc7c8fe27f1a80ae01edfacae391"} Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.472235 4857 scope.go:117] "RemoveContainer" containerID="e73de1d80b81f56a349fb898eac2fb806d6c36c416ab2aa788517d3eadadf750" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.472265 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wgwtn" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.473980 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-grx52" event={"ID":"e37b046b-176a-4bc3-bcbc-3093fe08b256","Type":"ContainerStarted","Data":"652f1ef9249db5801eb28e6063b4bbc1656b8490df0bb987de87871783cdf091"} Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.474246 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.477192 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m49jm" event={"ID":"c7894b0e-91be-4ed2-810b-f9e24df87fae","Type":"ContainerStarted","Data":"ff413035a44dde3b7d6ffcaa65f4490c335a60965b83096a6c6a0a9e9387a50e"} Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.480434 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" event={"ID":"0764bcb8-a3f2-46cd-bceb-f08b51012b7d","Type":"ContainerStarted","Data":"db400928b2dc43041c853c627caef7d633824f9f6fb73834d9675fbb566a8233"} Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.582425 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" podStartSLOduration=2.977713497 podStartE2EDuration="6.582406838s" podCreationTimestamp="2025-12-01 21:47:45 +0000 UTC" firstStartedPulling="2025-12-01 21:47:47.052001016 +0000 UTC m=+785.542063333" lastFinishedPulling="2025-12-01 21:47:50.656694357 +0000 UTC m=+789.146756674" observedRunningTime="2025-12-01 21:47:51.57930026 +0000 UTC m=+790.069362577" watchObservedRunningTime="2025-12-01 21:47:51.582406838 +0000 UTC m=+790.072469155" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.599035 4857 scope.go:117] "RemoveContainer" containerID="bda606175e68af92d9c75538f4bb325cf5e3e6792e9751ba83f4ce9be5dd2887" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.612429 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-78hdv" podStartSLOduration=2.504667973 podStartE2EDuration="6.612412113s" podCreationTimestamp="2025-12-01 21:47:45 +0000 UTC" firstStartedPulling="2025-12-01 21:47:46.52323695 +0000 UTC m=+785.013299267" lastFinishedPulling="2025-12-01 21:47:50.63098109 +0000 UTC m=+789.121043407" observedRunningTime="2025-12-01 21:47:51.597295762 +0000 UTC m=+790.087358099" watchObservedRunningTime="2025-12-01 21:47:51.612412113 +0000 UTC m=+790.102474430" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.615178 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-grx52" podStartSLOduration=2.461140217 podStartE2EDuration="6.615171172s" podCreationTimestamp="2025-12-01 21:47:45 +0000 UTC" firstStartedPulling="2025-12-01 21:47:46.504125269 +0000 UTC m=+784.994187576" lastFinishedPulling="2025-12-01 21:47:50.658156214 +0000 UTC m=+789.148218531" observedRunningTime="2025-12-01 21:47:51.61232803 +0000 UTC m=+790.102390367" watchObservedRunningTime="2025-12-01 21:47:51.615171172 +0000 UTC m=+790.105233489" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.630024 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wgwtn"] Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.633062 4857 scope.go:117] "RemoveContainer" containerID="2941770c7f5f647f417d319bb3728ce80d5f8612b48ed35941b8bf80fd065356" Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.640528 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wgwtn"] Dec 01 21:47:51 crc kubenswrapper[4857]: I1201 21:47:51.842281 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49e654be-8e0b-4421-8d29-c9708ba8af3e" path="/var/lib/kubelet/pods/49e654be-8e0b-4421-8d29-c9708ba8af3e/volumes" Dec 01 21:47:53 crc kubenswrapper[4857]: I1201 21:47:53.499695 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m49jm" event={"ID":"c7894b0e-91be-4ed2-810b-f9e24df87fae","Type":"ContainerStarted","Data":"3c9d4cf72829dadbdb8a98bc4d76f30f9f5e52f1debb9ff950f5a65b100270d2"} Dec 01 21:47:53 crc kubenswrapper[4857]: I1201 21:47:53.527461 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-m49jm" podStartSLOduration=2.053827752 podStartE2EDuration="8.527439408s" podCreationTimestamp="2025-12-01 21:47:45 +0000 UTC" firstStartedPulling="2025-12-01 21:47:46.687521174 +0000 UTC m=+785.177583511" lastFinishedPulling="2025-12-01 21:47:53.16113284 +0000 UTC m=+791.651195167" observedRunningTime="2025-12-01 21:47:53.52117328 +0000 UTC m=+792.011235657" watchObservedRunningTime="2025-12-01 21:47:53.527439408 +0000 UTC m=+792.017501735" Dec 01 21:47:56 crc kubenswrapper[4857]: I1201 21:47:56.530379 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:56 crc kubenswrapper[4857]: I1201 21:47:56.530887 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:56 crc kubenswrapper[4857]: I1201 21:47:56.530958 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-grx52" Dec 01 21:47:56 crc kubenswrapper[4857]: I1201 21:47:56.540717 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:57 crc kubenswrapper[4857]: I1201 21:47:57.576522 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-866cc8d79-t6gkb" Dec 01 21:47:57 crc kubenswrapper[4857]: I1201 21:47:57.687263 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-r5kwb"] Dec 01 21:47:57 crc kubenswrapper[4857]: I1201 21:47:57.806534 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:47:57 crc kubenswrapper[4857]: I1201 21:47:57.806609 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:48:06 crc kubenswrapper[4857]: I1201 21:48:06.770677 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-bdx6w" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.354308 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q"] Dec 01 21:48:22 crc kubenswrapper[4857]: E1201 21:48:22.355081 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e654be-8e0b-4421-8d29-c9708ba8af3e" containerName="registry-server" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.355096 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e654be-8e0b-4421-8d29-c9708ba8af3e" containerName="registry-server" Dec 01 21:48:22 crc kubenswrapper[4857]: E1201 21:48:22.355117 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e654be-8e0b-4421-8d29-c9708ba8af3e" containerName="extract-utilities" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.355123 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e654be-8e0b-4421-8d29-c9708ba8af3e" containerName="extract-utilities" Dec 01 21:48:22 crc kubenswrapper[4857]: E1201 21:48:22.355134 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e654be-8e0b-4421-8d29-c9708ba8af3e" containerName="extract-content" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.355139 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e654be-8e0b-4421-8d29-c9708ba8af3e" containerName="extract-content" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.355250 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e654be-8e0b-4421-8d29-c9708ba8af3e" containerName="registry-server" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.355964 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.357976 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.363289 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q"] Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.515965 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b421a90d-3d2c-48a5-b905-ed356dd54da2-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q\" (UID: \"b421a90d-3d2c-48a5-b905-ed356dd54da2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.516028 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b421a90d-3d2c-48a5-b905-ed356dd54da2-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q\" (UID: \"b421a90d-3d2c-48a5-b905-ed356dd54da2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.516075 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2lhl\" (UniqueName: \"kubernetes.io/projected/b421a90d-3d2c-48a5-b905-ed356dd54da2-kube-api-access-b2lhl\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q\" (UID: \"b421a90d-3d2c-48a5-b905-ed356dd54da2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.616858 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b421a90d-3d2c-48a5-b905-ed356dd54da2-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q\" (UID: \"b421a90d-3d2c-48a5-b905-ed356dd54da2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.617191 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b421a90d-3d2c-48a5-b905-ed356dd54da2-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q\" (UID: \"b421a90d-3d2c-48a5-b905-ed356dd54da2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.617344 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2lhl\" (UniqueName: \"kubernetes.io/projected/b421a90d-3d2c-48a5-b905-ed356dd54da2-kube-api-access-b2lhl\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q\" (UID: \"b421a90d-3d2c-48a5-b905-ed356dd54da2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.617519 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b421a90d-3d2c-48a5-b905-ed356dd54da2-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q\" (UID: \"b421a90d-3d2c-48a5-b905-ed356dd54da2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.617868 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b421a90d-3d2c-48a5-b905-ed356dd54da2-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q\" (UID: \"b421a90d-3d2c-48a5-b905-ed356dd54da2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.635645 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2lhl\" (UniqueName: \"kubernetes.io/projected/b421a90d-3d2c-48a5-b905-ed356dd54da2-kube-api-access-b2lhl\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q\" (UID: \"b421a90d-3d2c-48a5-b905-ed356dd54da2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.676832 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.751325 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-r5kwb" podUID="b8e5475f-0369-448a-895e-c02b78884677" containerName="console" containerID="cri-o://f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f" gracePeriod=15 Dec 01 21:48:22 crc kubenswrapper[4857]: I1201 21:48:22.954461 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q"] Dec 01 21:48:23 crc kubenswrapper[4857]: E1201 21:48:23.034561 4857 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8e5475f_0369_448a_895e_c02b78884677.slice/crio-conmon-f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f.scope\": RecentStats: unable to find data in memory cache]" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.746811 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-r5kwb_b8e5475f-0369-448a-895e-c02b78884677/console/0.log" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.747507 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.794413 4857 generic.go:334] "Generic (PLEG): container finished" podID="b421a90d-3d2c-48a5-b905-ed356dd54da2" containerID="e8929e99869be2782f4872536e31a80704c887a42e3b03cb92e91a6fec06a736" exitCode=0 Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.794522 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" event={"ID":"b421a90d-3d2c-48a5-b905-ed356dd54da2","Type":"ContainerDied","Data":"e8929e99869be2782f4872536e31a80704c887a42e3b03cb92e91a6fec06a736"} Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.794948 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" event={"ID":"b421a90d-3d2c-48a5-b905-ed356dd54da2","Type":"ContainerStarted","Data":"04499254e8af7fa7ea9266d76ca8d5faef2052856acf056870dc043dbd316d97"} Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.796259 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-r5kwb_b8e5475f-0369-448a-895e-c02b78884677/console/0.log" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.796308 4857 generic.go:334] "Generic (PLEG): container finished" podID="b8e5475f-0369-448a-895e-c02b78884677" containerID="f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f" exitCode=2 Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.796336 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r5kwb" event={"ID":"b8e5475f-0369-448a-895e-c02b78884677","Type":"ContainerDied","Data":"f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f"} Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.796363 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r5kwb" event={"ID":"b8e5475f-0369-448a-895e-c02b78884677","Type":"ContainerDied","Data":"4dc9776b7b82b01c9f92d944c9c0b1a1da61e7dda60172531293afe70af0ca3e"} Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.796384 4857 scope.go:117] "RemoveContainer" containerID="f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.796474 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r5kwb" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.822352 4857 scope.go:117] "RemoveContainer" containerID="f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f" Dec 01 21:48:23 crc kubenswrapper[4857]: E1201 21:48:23.822820 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f\": container with ID starting with f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f not found: ID does not exist" containerID="f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.822867 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f"} err="failed to get container status \"f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f\": rpc error: code = NotFound desc = could not find container \"f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f\": container with ID starting with f530be42c53db9dddf9ff742d49c7a723f3547e5f47ce133e1c280ea974d222f not found: ID does not exist" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.934030 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-service-ca\") pod \"b8e5475f-0369-448a-895e-c02b78884677\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.934142 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b8e5475f-0369-448a-895e-c02b78884677-console-serving-cert\") pod \"b8e5475f-0369-448a-895e-c02b78884677\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.934177 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b8e5475f-0369-448a-895e-c02b78884677-console-oauth-config\") pod \"b8e5475f-0369-448a-895e-c02b78884677\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.934195 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27hc5\" (UniqueName: \"kubernetes.io/projected/b8e5475f-0369-448a-895e-c02b78884677-kube-api-access-27hc5\") pod \"b8e5475f-0369-448a-895e-c02b78884677\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.934220 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-console-config\") pod \"b8e5475f-0369-448a-895e-c02b78884677\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.934262 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-oauth-serving-cert\") pod \"b8e5475f-0369-448a-895e-c02b78884677\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.934280 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-trusted-ca-bundle\") pod \"b8e5475f-0369-448a-895e-c02b78884677\" (UID: \"b8e5475f-0369-448a-895e-c02b78884677\") " Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.935074 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-service-ca" (OuterVolumeSpecName: "service-ca") pod "b8e5475f-0369-448a-895e-c02b78884677" (UID: "b8e5475f-0369-448a-895e-c02b78884677"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.935609 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-console-config" (OuterVolumeSpecName: "console-config") pod "b8e5475f-0369-448a-895e-c02b78884677" (UID: "b8e5475f-0369-448a-895e-c02b78884677"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.935740 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b8e5475f-0369-448a-895e-c02b78884677" (UID: "b8e5475f-0369-448a-895e-c02b78884677"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.936564 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "b8e5475f-0369-448a-895e-c02b78884677" (UID: "b8e5475f-0369-448a-895e-c02b78884677"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.941470 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8e5475f-0369-448a-895e-c02b78884677-kube-api-access-27hc5" (OuterVolumeSpecName: "kube-api-access-27hc5") pod "b8e5475f-0369-448a-895e-c02b78884677" (UID: "b8e5475f-0369-448a-895e-c02b78884677"). InnerVolumeSpecName "kube-api-access-27hc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.941921 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e5475f-0369-448a-895e-c02b78884677-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b8e5475f-0369-448a-895e-c02b78884677" (UID: "b8e5475f-0369-448a-895e-c02b78884677"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:48:23 crc kubenswrapper[4857]: I1201 21:48:23.948420 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e5475f-0369-448a-895e-c02b78884677-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b8e5475f-0369-448a-895e-c02b78884677" (UID: "b8e5475f-0369-448a-895e-c02b78884677"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:48:24 crc kubenswrapper[4857]: I1201 21:48:24.036068 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27hc5\" (UniqueName: \"kubernetes.io/projected/b8e5475f-0369-448a-895e-c02b78884677-kube-api-access-27hc5\") on node \"crc\" DevicePath \"\"" Dec 01 21:48:24 crc kubenswrapper[4857]: I1201 21:48:24.036129 4857 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b8e5475f-0369-448a-895e-c02b78884677-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:48:24 crc kubenswrapper[4857]: I1201 21:48:24.036149 4857 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:48:24 crc kubenswrapper[4857]: I1201 21:48:24.036166 4857 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:48:24 crc kubenswrapper[4857]: I1201 21:48:24.036185 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:48:24 crc kubenswrapper[4857]: I1201 21:48:24.036203 4857 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b8e5475f-0369-448a-895e-c02b78884677-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 21:48:24 crc kubenswrapper[4857]: I1201 21:48:24.036221 4857 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b8e5475f-0369-448a-895e-c02b78884677-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 21:48:24 crc kubenswrapper[4857]: I1201 21:48:24.162843 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-r5kwb"] Dec 01 21:48:24 crc kubenswrapper[4857]: I1201 21:48:24.176509 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-r5kwb"] Dec 01 21:48:25 crc kubenswrapper[4857]: I1201 21:48:25.851333 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8e5475f-0369-448a-895e-c02b78884677" path="/var/lib/kubelet/pods/b8e5475f-0369-448a-895e-c02b78884677/volumes" Dec 01 21:48:26 crc kubenswrapper[4857]: I1201 21:48:26.830999 4857 generic.go:334] "Generic (PLEG): container finished" podID="b421a90d-3d2c-48a5-b905-ed356dd54da2" containerID="d0932fee183cf3df4d779d7cd36f20de800d163dda24109d3a15a468b7dddab4" exitCode=0 Dec 01 21:48:26 crc kubenswrapper[4857]: I1201 21:48:26.831131 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" event={"ID":"b421a90d-3d2c-48a5-b905-ed356dd54da2","Type":"ContainerDied","Data":"d0932fee183cf3df4d779d7cd36f20de800d163dda24109d3a15a468b7dddab4"} Dec 01 21:48:27 crc kubenswrapper[4857]: I1201 21:48:27.806806 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:48:27 crc kubenswrapper[4857]: I1201 21:48:27.806908 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:48:27 crc kubenswrapper[4857]: I1201 21:48:27.842548 4857 generic.go:334] "Generic (PLEG): container finished" podID="b421a90d-3d2c-48a5-b905-ed356dd54da2" containerID="8ec21d168b13d64c4cf9d5737fd09e5e437d129b595665124c1fe95dd0f0d653" exitCode=0 Dec 01 21:48:27 crc kubenswrapper[4857]: I1201 21:48:27.847322 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" event={"ID":"b421a90d-3d2c-48a5-b905-ed356dd54da2","Type":"ContainerDied","Data":"8ec21d168b13d64c4cf9d5737fd09e5e437d129b595665124c1fe95dd0f0d653"} Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.252074 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.422767 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2lhl\" (UniqueName: \"kubernetes.io/projected/b421a90d-3d2c-48a5-b905-ed356dd54da2-kube-api-access-b2lhl\") pod \"b421a90d-3d2c-48a5-b905-ed356dd54da2\" (UID: \"b421a90d-3d2c-48a5-b905-ed356dd54da2\") " Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.423290 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b421a90d-3d2c-48a5-b905-ed356dd54da2-bundle\") pod \"b421a90d-3d2c-48a5-b905-ed356dd54da2\" (UID: \"b421a90d-3d2c-48a5-b905-ed356dd54da2\") " Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.423485 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b421a90d-3d2c-48a5-b905-ed356dd54da2-util\") pod \"b421a90d-3d2c-48a5-b905-ed356dd54da2\" (UID: \"b421a90d-3d2c-48a5-b905-ed356dd54da2\") " Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.424985 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b421a90d-3d2c-48a5-b905-ed356dd54da2-bundle" (OuterVolumeSpecName: "bundle") pod "b421a90d-3d2c-48a5-b905-ed356dd54da2" (UID: "b421a90d-3d2c-48a5-b905-ed356dd54da2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.431868 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b421a90d-3d2c-48a5-b905-ed356dd54da2-kube-api-access-b2lhl" (OuterVolumeSpecName: "kube-api-access-b2lhl") pod "b421a90d-3d2c-48a5-b905-ed356dd54da2" (UID: "b421a90d-3d2c-48a5-b905-ed356dd54da2"). InnerVolumeSpecName "kube-api-access-b2lhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.445390 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b421a90d-3d2c-48a5-b905-ed356dd54da2-util" (OuterVolumeSpecName: "util") pod "b421a90d-3d2c-48a5-b905-ed356dd54da2" (UID: "b421a90d-3d2c-48a5-b905-ed356dd54da2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.525015 4857 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b421a90d-3d2c-48a5-b905-ed356dd54da2-util\") on node \"crc\" DevicePath \"\"" Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.525277 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2lhl\" (UniqueName: \"kubernetes.io/projected/b421a90d-3d2c-48a5-b905-ed356dd54da2-kube-api-access-b2lhl\") on node \"crc\" DevicePath \"\"" Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.525384 4857 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b421a90d-3d2c-48a5-b905-ed356dd54da2-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.864025 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" event={"ID":"b421a90d-3d2c-48a5-b905-ed356dd54da2","Type":"ContainerDied","Data":"04499254e8af7fa7ea9266d76ca8d5faef2052856acf056870dc043dbd316d97"} Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.864103 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04499254e8af7fa7ea9266d76ca8d5faef2052856acf056870dc043dbd316d97" Dec 01 21:48:29 crc kubenswrapper[4857]: I1201 21:48:29.864707 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.686087 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx"] Dec 01 21:48:37 crc kubenswrapper[4857]: E1201 21:48:37.686807 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b421a90d-3d2c-48a5-b905-ed356dd54da2" containerName="pull" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.686819 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b421a90d-3d2c-48a5-b905-ed356dd54da2" containerName="pull" Dec 01 21:48:37 crc kubenswrapper[4857]: E1201 21:48:37.686831 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b421a90d-3d2c-48a5-b905-ed356dd54da2" containerName="util" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.686836 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b421a90d-3d2c-48a5-b905-ed356dd54da2" containerName="util" Dec 01 21:48:37 crc kubenswrapper[4857]: E1201 21:48:37.686851 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e5475f-0369-448a-895e-c02b78884677" containerName="console" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.686859 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e5475f-0369-448a-895e-c02b78884677" containerName="console" Dec 01 21:48:37 crc kubenswrapper[4857]: E1201 21:48:37.686867 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b421a90d-3d2c-48a5-b905-ed356dd54da2" containerName="extract" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.686872 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b421a90d-3d2c-48a5-b905-ed356dd54da2" containerName="extract" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.686964 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e5475f-0369-448a-895e-c02b78884677" containerName="console" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.686983 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b421a90d-3d2c-48a5-b905-ed356dd54da2" containerName="extract" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.687383 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.693734 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.693872 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.694204 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.694208 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-9rwwv" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.694679 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.710205 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx"] Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.855875 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhf5m\" (UniqueName: \"kubernetes.io/projected/8eee3586-4292-4165-81c4-496c2ee8ab9c-kube-api-access-lhf5m\") pod \"metallb-operator-controller-manager-75c9cd8c7-tt4cx\" (UID: \"8eee3586-4292-4165-81c4-496c2ee8ab9c\") " pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.855953 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8eee3586-4292-4165-81c4-496c2ee8ab9c-apiservice-cert\") pod \"metallb-operator-controller-manager-75c9cd8c7-tt4cx\" (UID: \"8eee3586-4292-4165-81c4-496c2ee8ab9c\") " pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.856394 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8eee3586-4292-4165-81c4-496c2ee8ab9c-webhook-cert\") pod \"metallb-operator-controller-manager-75c9cd8c7-tt4cx\" (UID: \"8eee3586-4292-4165-81c4-496c2ee8ab9c\") " pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.957620 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8eee3586-4292-4165-81c4-496c2ee8ab9c-apiservice-cert\") pod \"metallb-operator-controller-manager-75c9cd8c7-tt4cx\" (UID: \"8eee3586-4292-4165-81c4-496c2ee8ab9c\") " pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.957701 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8eee3586-4292-4165-81c4-496c2ee8ab9c-webhook-cert\") pod \"metallb-operator-controller-manager-75c9cd8c7-tt4cx\" (UID: \"8eee3586-4292-4165-81c4-496c2ee8ab9c\") " pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.957779 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhf5m\" (UniqueName: \"kubernetes.io/projected/8eee3586-4292-4165-81c4-496c2ee8ab9c-kube-api-access-lhf5m\") pod \"metallb-operator-controller-manager-75c9cd8c7-tt4cx\" (UID: \"8eee3586-4292-4165-81c4-496c2ee8ab9c\") " pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.967861 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8eee3586-4292-4165-81c4-496c2ee8ab9c-apiservice-cert\") pod \"metallb-operator-controller-manager-75c9cd8c7-tt4cx\" (UID: \"8eee3586-4292-4165-81c4-496c2ee8ab9c\") " pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.972537 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8eee3586-4292-4165-81c4-496c2ee8ab9c-webhook-cert\") pod \"metallb-operator-controller-manager-75c9cd8c7-tt4cx\" (UID: \"8eee3586-4292-4165-81c4-496c2ee8ab9c\") " pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:48:37 crc kubenswrapper[4857]: I1201 21:48:37.983653 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhf5m\" (UniqueName: \"kubernetes.io/projected/8eee3586-4292-4165-81c4-496c2ee8ab9c-kube-api-access-lhf5m\") pod \"metallb-operator-controller-manager-75c9cd8c7-tt4cx\" (UID: \"8eee3586-4292-4165-81c4-496c2ee8ab9c\") " pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.003820 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.057387 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs"] Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.058506 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.063673 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.063980 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-tndbv" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.064144 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.075270 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs"] Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.160101 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7hlm\" (UniqueName: \"kubernetes.io/projected/bd5afa07-4d43-4dce-8292-58ad63e6362a-kube-api-access-l7hlm\") pod \"metallb-operator-webhook-server-75b5f488d8-k8rxs\" (UID: \"bd5afa07-4d43-4dce-8292-58ad63e6362a\") " pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.160219 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bd5afa07-4d43-4dce-8292-58ad63e6362a-apiservice-cert\") pod \"metallb-operator-webhook-server-75b5f488d8-k8rxs\" (UID: \"bd5afa07-4d43-4dce-8292-58ad63e6362a\") " pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.160256 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bd5afa07-4d43-4dce-8292-58ad63e6362a-webhook-cert\") pod \"metallb-operator-webhook-server-75b5f488d8-k8rxs\" (UID: \"bd5afa07-4d43-4dce-8292-58ad63e6362a\") " pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.261807 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bd5afa07-4d43-4dce-8292-58ad63e6362a-apiservice-cert\") pod \"metallb-operator-webhook-server-75b5f488d8-k8rxs\" (UID: \"bd5afa07-4d43-4dce-8292-58ad63e6362a\") " pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.261863 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bd5afa07-4d43-4dce-8292-58ad63e6362a-webhook-cert\") pod \"metallb-operator-webhook-server-75b5f488d8-k8rxs\" (UID: \"bd5afa07-4d43-4dce-8292-58ad63e6362a\") " pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.261895 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7hlm\" (UniqueName: \"kubernetes.io/projected/bd5afa07-4d43-4dce-8292-58ad63e6362a-kube-api-access-l7hlm\") pod \"metallb-operator-webhook-server-75b5f488d8-k8rxs\" (UID: \"bd5afa07-4d43-4dce-8292-58ad63e6362a\") " pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.266292 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bd5afa07-4d43-4dce-8292-58ad63e6362a-apiservice-cert\") pod \"metallb-operator-webhook-server-75b5f488d8-k8rxs\" (UID: \"bd5afa07-4d43-4dce-8292-58ad63e6362a\") " pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.267389 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bd5afa07-4d43-4dce-8292-58ad63e6362a-webhook-cert\") pod \"metallb-operator-webhook-server-75b5f488d8-k8rxs\" (UID: \"bd5afa07-4d43-4dce-8292-58ad63e6362a\") " pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.283960 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx"] Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.289767 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7hlm\" (UniqueName: \"kubernetes.io/projected/bd5afa07-4d43-4dce-8292-58ad63e6362a-kube-api-access-l7hlm\") pod \"metallb-operator-webhook-server-75b5f488d8-k8rxs\" (UID: \"bd5afa07-4d43-4dce-8292-58ad63e6362a\") " pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:48:38 crc kubenswrapper[4857]: W1201 21:48:38.293332 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8eee3586_4292_4165_81c4_496c2ee8ab9c.slice/crio-06684e12f15b76a0c55c1d3d442c51299c42da69149bfcaeeab2335a3d19d409 WatchSource:0}: Error finding container 06684e12f15b76a0c55c1d3d442c51299c42da69149bfcaeeab2335a3d19d409: Status 404 returned error can't find the container with id 06684e12f15b76a0c55c1d3d442c51299c42da69149bfcaeeab2335a3d19d409 Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.426472 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.738605 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs"] Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.931447 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" event={"ID":"bd5afa07-4d43-4dce-8292-58ad63e6362a","Type":"ContainerStarted","Data":"c56cc35787acd8bd4ece56bd3dcea09d8b3eb75bda8ead6fd92d4343503121b9"} Dec 01 21:48:38 crc kubenswrapper[4857]: I1201 21:48:38.933062 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" event={"ID":"8eee3586-4292-4165-81c4-496c2ee8ab9c","Type":"ContainerStarted","Data":"06684e12f15b76a0c55c1d3d442c51299c42da69149bfcaeeab2335a3d19d409"} Dec 01 21:48:45 crc kubenswrapper[4857]: I1201 21:48:45.045482 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" event={"ID":"8eee3586-4292-4165-81c4-496c2ee8ab9c","Type":"ContainerStarted","Data":"ab9a841d49939c875a1674bb00bfde4c354fdcf6545c3b9e79707aebc1d905d2"} Dec 01 21:48:45 crc kubenswrapper[4857]: I1201 21:48:45.046159 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:48:45 crc kubenswrapper[4857]: I1201 21:48:45.073730 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" podStartSLOduration=1.539761771 podStartE2EDuration="8.073703475s" podCreationTimestamp="2025-12-01 21:48:37 +0000 UTC" firstStartedPulling="2025-12-01 21:48:38.297014055 +0000 UTC m=+836.787076372" lastFinishedPulling="2025-12-01 21:48:44.830955749 +0000 UTC m=+843.321018076" observedRunningTime="2025-12-01 21:48:45.071575696 +0000 UTC m=+843.561638023" watchObservedRunningTime="2025-12-01 21:48:45.073703475 +0000 UTC m=+843.563765832" Dec 01 21:48:46 crc kubenswrapper[4857]: I1201 21:48:46.055992 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" event={"ID":"bd5afa07-4d43-4dce-8292-58ad63e6362a","Type":"ContainerStarted","Data":"d8fc306221b81b45ee704efb699e1d6e64524fc82b07634123314e303fdd948b"} Dec 01 21:48:46 crc kubenswrapper[4857]: I1201 21:48:46.099225 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" podStartSLOduration=2.001862688 podStartE2EDuration="8.099204279s" podCreationTimestamp="2025-12-01 21:48:38 +0000 UTC" firstStartedPulling="2025-12-01 21:48:38.741251714 +0000 UTC m=+837.231314031" lastFinishedPulling="2025-12-01 21:48:44.838593305 +0000 UTC m=+843.328655622" observedRunningTime="2025-12-01 21:48:46.098797939 +0000 UTC m=+844.588860266" watchObservedRunningTime="2025-12-01 21:48:46.099204279 +0000 UTC m=+844.589266606" Dec 01 21:48:47 crc kubenswrapper[4857]: I1201 21:48:47.063732 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:48:57 crc kubenswrapper[4857]: I1201 21:48:57.806021 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:48:57 crc kubenswrapper[4857]: I1201 21:48:57.806495 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:48:57 crc kubenswrapper[4857]: I1201 21:48:57.806532 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:48:57 crc kubenswrapper[4857]: I1201 21:48:57.807083 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f890539e4830c07b0e2e556ca8ce83ae288e8de08c435fa380b5373247a6c2d8"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 21:48:57 crc kubenswrapper[4857]: I1201 21:48:57.807127 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://f890539e4830c07b0e2e556ca8ce83ae288e8de08c435fa380b5373247a6c2d8" gracePeriod=600 Dec 01 21:48:58 crc kubenswrapper[4857]: I1201 21:48:58.135148 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="f890539e4830c07b0e2e556ca8ce83ae288e8de08c435fa380b5373247a6c2d8" exitCode=0 Dec 01 21:48:58 crc kubenswrapper[4857]: I1201 21:48:58.135232 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"f890539e4830c07b0e2e556ca8ce83ae288e8de08c435fa380b5373247a6c2d8"} Dec 01 21:48:58 crc kubenswrapper[4857]: I1201 21:48:58.135489 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"1936180283cd1643268f62d5199bdc5d47710e031991bbd7a33aafc43d3ee862"} Dec 01 21:48:58 crc kubenswrapper[4857]: I1201 21:48:58.135515 4857 scope.go:117] "RemoveContainer" containerID="697c28f6400e4a41370ca0b263896ab5c011bd878a53e6f826180b346cdc9317" Dec 01 21:48:58 crc kubenswrapper[4857]: I1201 21:48:58.433574 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-75b5f488d8-k8rxs" Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.283574 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nm4pl"] Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.286037 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.295012 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nm4pl"] Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.396120 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czctv\" (UniqueName: \"kubernetes.io/projected/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-kube-api-access-czctv\") pod \"community-operators-nm4pl\" (UID: \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\") " pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.396194 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-catalog-content\") pod \"community-operators-nm4pl\" (UID: \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\") " pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.396223 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-utilities\") pod \"community-operators-nm4pl\" (UID: \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\") " pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.497121 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-catalog-content\") pod \"community-operators-nm4pl\" (UID: \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\") " pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.497480 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-utilities\") pod \"community-operators-nm4pl\" (UID: \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\") " pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.497534 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czctv\" (UniqueName: \"kubernetes.io/projected/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-kube-api-access-czctv\") pod \"community-operators-nm4pl\" (UID: \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\") " pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.497916 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-catalog-content\") pod \"community-operators-nm4pl\" (UID: \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\") " pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.497960 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-utilities\") pod \"community-operators-nm4pl\" (UID: \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\") " pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.525451 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czctv\" (UniqueName: \"kubernetes.io/projected/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-kube-api-access-czctv\") pod \"community-operators-nm4pl\" (UID: \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\") " pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:13 crc kubenswrapper[4857]: I1201 21:49:13.622448 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:14 crc kubenswrapper[4857]: I1201 21:49:14.160251 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nm4pl"] Dec 01 21:49:14 crc kubenswrapper[4857]: W1201 21:49:14.172281 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75e0d3a3_89e2_41bb_b74d_f0a49b795b72.slice/crio-045bf66f0d3e59b26928fc6084812a1f407813ec9177bf612b0b912f832ea8f7 WatchSource:0}: Error finding container 045bf66f0d3e59b26928fc6084812a1f407813ec9177bf612b0b912f832ea8f7: Status 404 returned error can't find the container with id 045bf66f0d3e59b26928fc6084812a1f407813ec9177bf612b0b912f832ea8f7 Dec 01 21:49:14 crc kubenswrapper[4857]: I1201 21:49:14.271901 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nm4pl" event={"ID":"75e0d3a3-89e2-41bb-b74d-f0a49b795b72","Type":"ContainerStarted","Data":"045bf66f0d3e59b26928fc6084812a1f407813ec9177bf612b0b912f832ea8f7"} Dec 01 21:49:15 crc kubenswrapper[4857]: I1201 21:49:15.278737 4857 generic.go:334] "Generic (PLEG): container finished" podID="75e0d3a3-89e2-41bb-b74d-f0a49b795b72" containerID="c48f755fd0040cde85a04f09030be7fc1afcdce53848671ea16789e61d27382b" exitCode=0 Dec 01 21:49:15 crc kubenswrapper[4857]: I1201 21:49:15.278803 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nm4pl" event={"ID":"75e0d3a3-89e2-41bb-b74d-f0a49b795b72","Type":"ContainerDied","Data":"c48f755fd0040cde85a04f09030be7fc1afcdce53848671ea16789e61d27382b"} Dec 01 21:49:17 crc kubenswrapper[4857]: I1201 21:49:17.299037 4857 generic.go:334] "Generic (PLEG): container finished" podID="75e0d3a3-89e2-41bb-b74d-f0a49b795b72" containerID="5024d2f7c4c776af696bde21d719f5f120bf68a70006b4d55e052712531b9353" exitCode=0 Dec 01 21:49:17 crc kubenswrapper[4857]: I1201 21:49:17.299262 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nm4pl" event={"ID":"75e0d3a3-89e2-41bb-b74d-f0a49b795b72","Type":"ContainerDied","Data":"5024d2f7c4c776af696bde21d719f5f120bf68a70006b4d55e052712531b9353"} Dec 01 21:49:18 crc kubenswrapper[4857]: I1201 21:49:18.011528 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-75c9cd8c7-tt4cx" Dec 01 21:49:18 crc kubenswrapper[4857]: I1201 21:49:18.949295 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-wjx7t"] Dec 01 21:49:18 crc kubenswrapper[4857]: I1201 21:49:18.961799 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:18 crc kubenswrapper[4857]: I1201 21:49:18.966358 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 01 21:49:18 crc kubenswrapper[4857]: I1201 21:49:18.967599 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 01 21:49:18 crc kubenswrapper[4857]: I1201 21:49:18.967738 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-x2vb5" Dec 01 21:49:18 crc kubenswrapper[4857]: I1201 21:49:18.982670 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn"] Dec 01 21:49:18 crc kubenswrapper[4857]: I1201 21:49:18.984001 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" Dec 01 21:49:18 crc kubenswrapper[4857]: I1201 21:49:18.993918 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/994ed59f-d2c6-4af0-b868-f167195e071c-frr-sockets\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:18 crc kubenswrapper[4857]: I1201 21:49:18.994009 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/994ed59f-d2c6-4af0-b868-f167195e071c-frr-startup\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.004470 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn"] Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.009222 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.044554 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-r64qq"] Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.045887 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.051122 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-85f5q" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.051293 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.051427 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.051532 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.065898 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-2gzlv"] Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.067097 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.069235 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.077820 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-2gzlv"] Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.113394 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/994ed59f-d2c6-4af0-b868-f167195e071c-metrics\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.113441 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5plcg\" (UniqueName: \"kubernetes.io/projected/994ed59f-d2c6-4af0-b868-f167195e071c-kube-api-access-5plcg\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.113512 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/994ed59f-d2c6-4af0-b868-f167195e071c-frr-sockets\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.113828 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/994ed59f-d2c6-4af0-b868-f167195e071c-frr-conf\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.113978 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/994ed59f-d2c6-4af0-b868-f167195e071c-metrics-certs\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.114215 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee319d44-c1a3-4325-835f-3d9fc6a187fd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-x6nqn\" (UID: \"ee319d44-c1a3-4325-835f-3d9fc6a187fd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.114265 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/994ed59f-d2c6-4af0-b868-f167195e071c-frr-sockets\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.114454 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/994ed59f-d2c6-4af0-b868-f167195e071c-frr-startup\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.114526 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/994ed59f-d2c6-4af0-b868-f167195e071c-reloader\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.114585 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4wtt\" (UniqueName: \"kubernetes.io/projected/ee319d44-c1a3-4325-835f-3d9fc6a187fd-kube-api-access-z4wtt\") pod \"frr-k8s-webhook-server-7fcb986d4-x6nqn\" (UID: \"ee319d44-c1a3-4325-835f-3d9fc6a187fd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.115884 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/994ed59f-d2c6-4af0-b868-f167195e071c-frr-startup\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.215696 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/994ed59f-d2c6-4af0-b868-f167195e071c-metrics\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.215747 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5plcg\" (UniqueName: \"kubernetes.io/projected/994ed59f-d2c6-4af0-b868-f167195e071c-kube-api-access-5plcg\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.215786 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-memberlist\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.215815 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/994ed59f-d2c6-4af0-b868-f167195e071c-frr-conf\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.215834 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-metrics-certs\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.215849 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/994ed59f-d2c6-4af0-b868-f167195e071c-metrics-certs\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.215870 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee319d44-c1a3-4325-835f-3d9fc6a187fd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-x6nqn\" (UID: \"ee319d44-c1a3-4325-835f-3d9fc6a187fd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.215889 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4-cert\") pod \"controller-f8648f98b-2gzlv\" (UID: \"28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4\") " pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.215905 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4-metrics-certs\") pod \"controller-f8648f98b-2gzlv\" (UID: \"28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4\") " pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.215950 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/994ed59f-d2c6-4af0-b868-f167195e071c-reloader\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.215993 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lqdz\" (UniqueName: \"kubernetes.io/projected/c6620f92-fe9f-4dde-a758-b46475bd8db8-kube-api-access-8lqdz\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.216020 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4wtt\" (UniqueName: \"kubernetes.io/projected/ee319d44-c1a3-4325-835f-3d9fc6a187fd-kube-api-access-z4wtt\") pod \"frr-k8s-webhook-server-7fcb986d4-x6nqn\" (UID: \"ee319d44-c1a3-4325-835f-3d9fc6a187fd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.216064 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cmj6\" (UniqueName: \"kubernetes.io/projected/28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4-kube-api-access-9cmj6\") pod \"controller-f8648f98b-2gzlv\" (UID: \"28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4\") " pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.216087 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c6620f92-fe9f-4dde-a758-b46475bd8db8-metallb-excludel2\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.216433 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/994ed59f-d2c6-4af0-b868-f167195e071c-metrics\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.216875 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/994ed59f-d2c6-4af0-b868-f167195e071c-frr-conf\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.218183 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/994ed59f-d2c6-4af0-b868-f167195e071c-reloader\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.223198 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee319d44-c1a3-4325-835f-3d9fc6a187fd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-x6nqn\" (UID: \"ee319d44-c1a3-4325-835f-3d9fc6a187fd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.225493 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/994ed59f-d2c6-4af0-b868-f167195e071c-metrics-certs\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.234834 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5plcg\" (UniqueName: \"kubernetes.io/projected/994ed59f-d2c6-4af0-b868-f167195e071c-kube-api-access-5plcg\") pod \"frr-k8s-wjx7t\" (UID: \"994ed59f-d2c6-4af0-b868-f167195e071c\") " pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.245750 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4wtt\" (UniqueName: \"kubernetes.io/projected/ee319d44-c1a3-4325-835f-3d9fc6a187fd-kube-api-access-z4wtt\") pod \"frr-k8s-webhook-server-7fcb986d4-x6nqn\" (UID: \"ee319d44-c1a3-4325-835f-3d9fc6a187fd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.310599 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.315708 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nm4pl" event={"ID":"75e0d3a3-89e2-41bb-b74d-f0a49b795b72","Type":"ContainerStarted","Data":"168db23371f32ac38840df8fb2b057e0bac7f99528d63d4783188fe133ae1906"} Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.317110 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-memberlist\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.317158 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-metrics-certs\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: E1201 21:49:19.317278 4857 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 21:49:19 crc kubenswrapper[4857]: E1201 21:49:19.317332 4857 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 01 21:49:19 crc kubenswrapper[4857]: E1201 21:49:19.317341 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-memberlist podName:c6620f92-fe9f-4dde-a758-b46475bd8db8 nodeName:}" failed. No retries permitted until 2025-12-01 21:49:19.817325738 +0000 UTC m=+878.307388055 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-memberlist") pod "speaker-r64qq" (UID: "c6620f92-fe9f-4dde-a758-b46475bd8db8") : secret "metallb-memberlist" not found Dec 01 21:49:19 crc kubenswrapper[4857]: E1201 21:49:19.317397 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-metrics-certs podName:c6620f92-fe9f-4dde-a758-b46475bd8db8 nodeName:}" failed. No retries permitted until 2025-12-01 21:49:19.817379489 +0000 UTC m=+878.307441896 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-metrics-certs") pod "speaker-r64qq" (UID: "c6620f92-fe9f-4dde-a758-b46475bd8db8") : secret "speaker-certs-secret" not found Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.317422 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4-cert\") pod \"controller-f8648f98b-2gzlv\" (UID: \"28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4\") " pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.317448 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4-metrics-certs\") pod \"controller-f8648f98b-2gzlv\" (UID: \"28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4\") " pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.317694 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lqdz\" (UniqueName: \"kubernetes.io/projected/c6620f92-fe9f-4dde-a758-b46475bd8db8-kube-api-access-8lqdz\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.317729 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c6620f92-fe9f-4dde-a758-b46475bd8db8-metallb-excludel2\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.317752 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cmj6\" (UniqueName: \"kubernetes.io/projected/28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4-kube-api-access-9cmj6\") pod \"controller-f8648f98b-2gzlv\" (UID: \"28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4\") " pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.318812 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c6620f92-fe9f-4dde-a758-b46475bd8db8-metallb-excludel2\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.320023 4857 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.328106 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.336345 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4-metrics-certs\") pod \"controller-f8648f98b-2gzlv\" (UID: \"28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4\") " pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.349368 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4-cert\") pod \"controller-f8648f98b-2gzlv\" (UID: \"28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4\") " pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.349750 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nm4pl" podStartSLOduration=3.47826753 podStartE2EDuration="6.349740942s" podCreationTimestamp="2025-12-01 21:49:13 +0000 UTC" firstStartedPulling="2025-12-01 21:49:15.28006369 +0000 UTC m=+873.770126007" lastFinishedPulling="2025-12-01 21:49:18.151537092 +0000 UTC m=+876.641599419" observedRunningTime="2025-12-01 21:49:19.347517641 +0000 UTC m=+877.837579968" watchObservedRunningTime="2025-12-01 21:49:19.349740942 +0000 UTC m=+877.839803259" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.350496 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cmj6\" (UniqueName: \"kubernetes.io/projected/28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4-kube-api-access-9cmj6\") pod \"controller-f8648f98b-2gzlv\" (UID: \"28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4\") " pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.355307 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lqdz\" (UniqueName: \"kubernetes.io/projected/c6620f92-fe9f-4dde-a758-b46475bd8db8-kube-api-access-8lqdz\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.382415 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.824960 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-memberlist\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.825363 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-metrics-certs\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: E1201 21:49:19.825173 4857 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 21:49:19 crc kubenswrapper[4857]: E1201 21:49:19.825961 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-memberlist podName:c6620f92-fe9f-4dde-a758-b46475bd8db8 nodeName:}" failed. No retries permitted until 2025-12-01 21:49:20.82594271 +0000 UTC m=+879.316005027 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-memberlist") pod "speaker-r64qq" (UID: "c6620f92-fe9f-4dde-a758-b46475bd8db8") : secret "metallb-memberlist" not found Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.836100 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-metrics-certs\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:19 crc kubenswrapper[4857]: I1201 21:49:19.958833 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn"] Dec 01 21:49:19 crc kubenswrapper[4857]: W1201 21:49:19.963446 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee319d44_c1a3_4325_835f_3d9fc6a187fd.slice/crio-611d3940919f592a5a83161db93b63b38d9d1ee3ef77f2f951e88060da7f7499 WatchSource:0}: Error finding container 611d3940919f592a5a83161db93b63b38d9d1ee3ef77f2f951e88060da7f7499: Status 404 returned error can't find the container with id 611d3940919f592a5a83161db93b63b38d9d1ee3ef77f2f951e88060da7f7499 Dec 01 21:49:20 crc kubenswrapper[4857]: I1201 21:49:20.090908 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-2gzlv"] Dec 01 21:49:20 crc kubenswrapper[4857]: I1201 21:49:20.322605 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjx7t" event={"ID":"994ed59f-d2c6-4af0-b868-f167195e071c","Type":"ContainerStarted","Data":"ad493fbbc3521b7f4a2d1f62d1301a460dfebcaa00d1beb9d8c0d0418a396fbd"} Dec 01 21:49:20 crc kubenswrapper[4857]: I1201 21:49:20.323715 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" event={"ID":"ee319d44-c1a3-4325-835f-3d9fc6a187fd","Type":"ContainerStarted","Data":"611d3940919f592a5a83161db93b63b38d9d1ee3ef77f2f951e88060da7f7499"} Dec 01 21:49:20 crc kubenswrapper[4857]: I1201 21:49:20.325381 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-2gzlv" event={"ID":"28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4","Type":"ContainerStarted","Data":"8a20981adb53df9a697f1fdda939cd277994950d2350720924b1325472f4fcda"} Dec 01 21:49:20 crc kubenswrapper[4857]: I1201 21:49:20.325432 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-2gzlv" event={"ID":"28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4","Type":"ContainerStarted","Data":"5e3878eb2d400a406ded407321565f30fa1b5a993d242aa4d3a43f662e5d124b"} Dec 01 21:49:20 crc kubenswrapper[4857]: I1201 21:49:20.839917 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-memberlist\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:20 crc kubenswrapper[4857]: I1201 21:49:20.844536 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c6620f92-fe9f-4dde-a758-b46475bd8db8-memberlist\") pod \"speaker-r64qq\" (UID: \"c6620f92-fe9f-4dde-a758-b46475bd8db8\") " pod="metallb-system/speaker-r64qq" Dec 01 21:49:20 crc kubenswrapper[4857]: I1201 21:49:20.867076 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-r64qq" Dec 01 21:49:20 crc kubenswrapper[4857]: W1201 21:49:20.909458 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6620f92_fe9f_4dde_a758_b46475bd8db8.slice/crio-20a5524e30a997dc37bfb3b68e67e98ab35c0d26fa1ec32f5b7d8fc5f08d08c1 WatchSource:0}: Error finding container 20a5524e30a997dc37bfb3b68e67e98ab35c0d26fa1ec32f5b7d8fc5f08d08c1: Status 404 returned error can't find the container with id 20a5524e30a997dc37bfb3b68e67e98ab35c0d26fa1ec32f5b7d8fc5f08d08c1 Dec 01 21:49:21 crc kubenswrapper[4857]: I1201 21:49:21.338224 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-2gzlv" event={"ID":"28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4","Type":"ContainerStarted","Data":"a835167406c6f3ec49c816aab2fc52b0eac963211a3ae09a0d6c5e2a5e3828db"} Dec 01 21:49:21 crc kubenswrapper[4857]: I1201 21:49:21.338627 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:21 crc kubenswrapper[4857]: I1201 21:49:21.341521 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r64qq" event={"ID":"c6620f92-fe9f-4dde-a758-b46475bd8db8","Type":"ContainerStarted","Data":"c46e154983d60e083e9dd399d0815d31022f034b59bcb063d3782482d34785f3"} Dec 01 21:49:21 crc kubenswrapper[4857]: I1201 21:49:21.341570 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r64qq" event={"ID":"c6620f92-fe9f-4dde-a758-b46475bd8db8","Type":"ContainerStarted","Data":"20a5524e30a997dc37bfb3b68e67e98ab35c0d26fa1ec32f5b7d8fc5f08d08c1"} Dec 01 21:49:21 crc kubenswrapper[4857]: I1201 21:49:21.935130 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-2gzlv" podStartSLOduration=2.935101142 podStartE2EDuration="2.935101142s" podCreationTimestamp="2025-12-01 21:49:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:49:21.35724704 +0000 UTC m=+879.847309357" watchObservedRunningTime="2025-12-01 21:49:21.935101142 +0000 UTC m=+880.425163459" Dec 01 21:49:22 crc kubenswrapper[4857]: I1201 21:49:22.366396 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r64qq" event={"ID":"c6620f92-fe9f-4dde-a758-b46475bd8db8","Type":"ContainerStarted","Data":"9e99cedca4bfb5cb70197e7bdeef27756fc3130e63c352bfa0b0bcad85339fa0"} Dec 01 21:49:22 crc kubenswrapper[4857]: I1201 21:49:22.367550 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-r64qq" Dec 01 21:49:23 crc kubenswrapper[4857]: I1201 21:49:23.623544 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:23 crc kubenswrapper[4857]: I1201 21:49:23.623595 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:23 crc kubenswrapper[4857]: I1201 21:49:23.682837 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:23 crc kubenswrapper[4857]: I1201 21:49:23.702855 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-r64qq" podStartSLOduration=4.7028386730000005 podStartE2EDuration="4.702838673s" podCreationTimestamp="2025-12-01 21:49:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:49:22.401642767 +0000 UTC m=+880.891705084" watchObservedRunningTime="2025-12-01 21:49:23.702838673 +0000 UTC m=+882.192900990" Dec 01 21:49:24 crc kubenswrapper[4857]: I1201 21:49:24.645391 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:24 crc kubenswrapper[4857]: I1201 21:49:24.698006 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nm4pl"] Dec 01 21:49:26 crc kubenswrapper[4857]: I1201 21:49:26.425361 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nm4pl" podUID="75e0d3a3-89e2-41bb-b74d-f0a49b795b72" containerName="registry-server" containerID="cri-o://168db23371f32ac38840df8fb2b057e0bac7f99528d63d4783188fe133ae1906" gracePeriod=2 Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.015517 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ql4mw"] Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.017481 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.033161 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ql4mw"] Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.212321 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-utilities\") pod \"redhat-marketplace-ql4mw\" (UID: \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\") " pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.212529 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-catalog-content\") pod \"redhat-marketplace-ql4mw\" (UID: \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\") " pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.212603 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjd2n\" (UniqueName: \"kubernetes.io/projected/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-kube-api-access-cjd2n\") pod \"redhat-marketplace-ql4mw\" (UID: \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\") " pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.313786 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-catalog-content\") pod \"redhat-marketplace-ql4mw\" (UID: \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\") " pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.313860 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjd2n\" (UniqueName: \"kubernetes.io/projected/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-kube-api-access-cjd2n\") pod \"redhat-marketplace-ql4mw\" (UID: \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\") " pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.313899 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-utilities\") pod \"redhat-marketplace-ql4mw\" (UID: \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\") " pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.314637 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-catalog-content\") pod \"redhat-marketplace-ql4mw\" (UID: \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\") " pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.314710 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-utilities\") pod \"redhat-marketplace-ql4mw\" (UID: \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\") " pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.337931 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjd2n\" (UniqueName: \"kubernetes.io/projected/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-kube-api-access-cjd2n\") pod \"redhat-marketplace-ql4mw\" (UID: \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\") " pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.351110 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.444611 4857 generic.go:334] "Generic (PLEG): container finished" podID="75e0d3a3-89e2-41bb-b74d-f0a49b795b72" containerID="168db23371f32ac38840df8fb2b057e0bac7f99528d63d4783188fe133ae1906" exitCode=0 Dec 01 21:49:27 crc kubenswrapper[4857]: I1201 21:49:27.444655 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nm4pl" event={"ID":"75e0d3a3-89e2-41bb-b74d-f0a49b795b72","Type":"ContainerDied","Data":"168db23371f32ac38840df8fb2b057e0bac7f99528d63d4783188fe133ae1906"} Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.028026 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ql4mw"] Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.029463 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:31 crc kubenswrapper[4857]: W1201 21:49:31.038743 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf78ae4f8_29a7_4ac1_b90d_cef7983c0a53.slice/crio-3b02b462bb85d9ed51baeeba2c31aec7b22414b1d506bdd0c3245fdce87ffcfa WatchSource:0}: Error finding container 3b02b462bb85d9ed51baeeba2c31aec7b22414b1d506bdd0c3245fdce87ffcfa: Status 404 returned error can't find the container with id 3b02b462bb85d9ed51baeeba2c31aec7b22414b1d506bdd0c3245fdce87ffcfa Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.044199 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-catalog-content\") pod \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\" (UID: \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\") " Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.044355 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-utilities\") pod \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\" (UID: \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\") " Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.044384 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czctv\" (UniqueName: \"kubernetes.io/projected/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-kube-api-access-czctv\") pod \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\" (UID: \"75e0d3a3-89e2-41bb-b74d-f0a49b795b72\") " Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.045776 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-utilities" (OuterVolumeSpecName: "utilities") pod "75e0d3a3-89e2-41bb-b74d-f0a49b795b72" (UID: "75e0d3a3-89e2-41bb-b74d-f0a49b795b72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.052782 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-kube-api-access-czctv" (OuterVolumeSpecName: "kube-api-access-czctv") pod "75e0d3a3-89e2-41bb-b74d-f0a49b795b72" (UID: "75e0d3a3-89e2-41bb-b74d-f0a49b795b72"). InnerVolumeSpecName "kube-api-access-czctv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.106659 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75e0d3a3-89e2-41bb-b74d-f0a49b795b72" (UID: "75e0d3a3-89e2-41bb-b74d-f0a49b795b72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.145179 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.145212 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.145222 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czctv\" (UniqueName: \"kubernetes.io/projected/75e0d3a3-89e2-41bb-b74d-f0a49b795b72-kube-api-access-czctv\") on node \"crc\" DevicePath \"\"" Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.536031 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nm4pl" event={"ID":"75e0d3a3-89e2-41bb-b74d-f0a49b795b72","Type":"ContainerDied","Data":"045bf66f0d3e59b26928fc6084812a1f407813ec9177bf612b0b912f832ea8f7"} Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.536447 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nm4pl" Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.536577 4857 scope.go:117] "RemoveContainer" containerID="168db23371f32ac38840df8fb2b057e0bac7f99528d63d4783188fe133ae1906" Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.540439 4857 generic.go:334] "Generic (PLEG): container finished" podID="f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" containerID="fa4e8f43e6c2359377df6a7d9bdb8eaa0cc2a5336d6290bcb66b877f22d654fe" exitCode=0 Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.540551 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ql4mw" event={"ID":"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53","Type":"ContainerDied","Data":"fa4e8f43e6c2359377df6a7d9bdb8eaa0cc2a5336d6290bcb66b877f22d654fe"} Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.540583 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ql4mw" event={"ID":"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53","Type":"ContainerStarted","Data":"3b02b462bb85d9ed51baeeba2c31aec7b22414b1d506bdd0c3245fdce87ffcfa"} Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.546576 4857 generic.go:334] "Generic (PLEG): container finished" podID="994ed59f-d2c6-4af0-b868-f167195e071c" containerID="a02e2eea24600ba30220011623cef82cd3b13f860f25823a449006d5574f8efa" exitCode=0 Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.546680 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjx7t" event={"ID":"994ed59f-d2c6-4af0-b868-f167195e071c","Type":"ContainerDied","Data":"a02e2eea24600ba30220011623cef82cd3b13f860f25823a449006d5574f8efa"} Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.551348 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" event={"ID":"ee319d44-c1a3-4325-835f-3d9fc6a187fd","Type":"ContainerStarted","Data":"87e68631c25b432e29b1ec9e9c3857c92e0a8fa12088e26f56acab22c2a7917d"} Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.552275 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.563818 4857 scope.go:117] "RemoveContainer" containerID="5024d2f7c4c776af696bde21d719f5f120bf68a70006b4d55e052712531b9353" Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.608644 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" podStartSLOduration=2.7410909180000003 podStartE2EDuration="13.60860618s" podCreationTimestamp="2025-12-01 21:49:18 +0000 UTC" firstStartedPulling="2025-12-01 21:49:19.966624961 +0000 UTC m=+878.456687278" lastFinishedPulling="2025-12-01 21:49:30.834140223 +0000 UTC m=+889.324202540" observedRunningTime="2025-12-01 21:49:31.59640178 +0000 UTC m=+890.086464107" watchObservedRunningTime="2025-12-01 21:49:31.60860618 +0000 UTC m=+890.098668547" Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.620442 4857 scope.go:117] "RemoveContainer" containerID="c48f755fd0040cde85a04f09030be7fc1afcdce53848671ea16789e61d27382b" Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.789160 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nm4pl"] Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.793074 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nm4pl"] Dec 01 21:49:31 crc kubenswrapper[4857]: I1201 21:49:31.841034 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75e0d3a3-89e2-41bb-b74d-f0a49b795b72" path="/var/lib/kubelet/pods/75e0d3a3-89e2-41bb-b74d-f0a49b795b72/volumes" Dec 01 21:49:32 crc kubenswrapper[4857]: I1201 21:49:32.577257 4857 generic.go:334] "Generic (PLEG): container finished" podID="994ed59f-d2c6-4af0-b868-f167195e071c" containerID="0d35484b8bf65884680c691b42d392bc4f1d0764280955ea8be68ec788da5adf" exitCode=0 Dec 01 21:49:32 crc kubenswrapper[4857]: I1201 21:49:32.577601 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjx7t" event={"ID":"994ed59f-d2c6-4af0-b868-f167195e071c","Type":"ContainerDied","Data":"0d35484b8bf65884680c691b42d392bc4f1d0764280955ea8be68ec788da5adf"} Dec 01 21:49:33 crc kubenswrapper[4857]: I1201 21:49:33.591945 4857 generic.go:334] "Generic (PLEG): container finished" podID="994ed59f-d2c6-4af0-b868-f167195e071c" containerID="de293ed5a714a08e7329bce9f46e567bba6cda60f4cae286dd51725e7263eec0" exitCode=0 Dec 01 21:49:33 crc kubenswrapper[4857]: I1201 21:49:33.592091 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjx7t" event={"ID":"994ed59f-d2c6-4af0-b868-f167195e071c","Type":"ContainerDied","Data":"de293ed5a714a08e7329bce9f46e567bba6cda60f4cae286dd51725e7263eec0"} Dec 01 21:49:33 crc kubenswrapper[4857]: I1201 21:49:33.594587 4857 generic.go:334] "Generic (PLEG): container finished" podID="f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" containerID="797c6aae26d02cfa40a2f5cd1cb6bd3bb06bbb35e097f45e76fc4b03796677fc" exitCode=0 Dec 01 21:49:33 crc kubenswrapper[4857]: I1201 21:49:33.594676 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ql4mw" event={"ID":"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53","Type":"ContainerDied","Data":"797c6aae26d02cfa40a2f5cd1cb6bd3bb06bbb35e097f45e76fc4b03796677fc"} Dec 01 21:49:34 crc kubenswrapper[4857]: I1201 21:49:34.605764 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjx7t" event={"ID":"994ed59f-d2c6-4af0-b868-f167195e071c","Type":"ContainerStarted","Data":"a877305b2f1315692dc8f00cbe24a91667470a40883f7aac40497853ac6cc39c"} Dec 01 21:49:34 crc kubenswrapper[4857]: I1201 21:49:34.606313 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjx7t" event={"ID":"994ed59f-d2c6-4af0-b868-f167195e071c","Type":"ContainerStarted","Data":"bf9bfaa2e9e886f2f84c4911b7102c295be17179801e167657368c09697c2362"} Dec 01 21:49:34 crc kubenswrapper[4857]: I1201 21:49:34.606324 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjx7t" event={"ID":"994ed59f-d2c6-4af0-b868-f167195e071c","Type":"ContainerStarted","Data":"bd6adfe3de5dfe68c01a701efae5d106ccc935381cad0d9adb514bcc20319148"} Dec 01 21:49:34 crc kubenswrapper[4857]: I1201 21:49:34.606332 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjx7t" event={"ID":"994ed59f-d2c6-4af0-b868-f167195e071c","Type":"ContainerStarted","Data":"e55cce40dd40705ad2f1c595f06fe091b8a21e5349a30b069b44e0da5edf3077"} Dec 01 21:49:34 crc kubenswrapper[4857]: I1201 21:49:34.606342 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjx7t" event={"ID":"994ed59f-d2c6-4af0-b868-f167195e071c","Type":"ContainerStarted","Data":"8ea1874474574bbf35c8e99e61d40abd576ed90572e194fa6fe7be72ba78b28f"} Dec 01 21:49:34 crc kubenswrapper[4857]: I1201 21:49:34.610224 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ql4mw" event={"ID":"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53","Type":"ContainerStarted","Data":"f4fb541a41d4163d50a9319cced538a8c3aec50fb1428dd1c3c4c9989246c7ed"} Dec 01 21:49:34 crc kubenswrapper[4857]: I1201 21:49:34.638804 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ql4mw" podStartSLOduration=6.010280016 podStartE2EDuration="8.638789818s" podCreationTimestamp="2025-12-01 21:49:26 +0000 UTC" firstStartedPulling="2025-12-01 21:49:31.54631208 +0000 UTC m=+890.036374437" lastFinishedPulling="2025-12-01 21:49:34.174821922 +0000 UTC m=+892.664884239" observedRunningTime="2025-12-01 21:49:34.634537151 +0000 UTC m=+893.124599468" watchObservedRunningTime="2025-12-01 21:49:34.638789818 +0000 UTC m=+893.128852135" Dec 01 21:49:35 crc kubenswrapper[4857]: I1201 21:49:35.622456 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjx7t" event={"ID":"994ed59f-d2c6-4af0-b868-f167195e071c","Type":"ContainerStarted","Data":"1b27ecea4c6e7274f1d4e12dd70fbc02c2602fb928e29baa7c1a0190960553a2"} Dec 01 21:49:35 crc kubenswrapper[4857]: I1201 21:49:35.622850 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:37 crc kubenswrapper[4857]: I1201 21:49:37.351578 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:37 crc kubenswrapper[4857]: I1201 21:49:37.352121 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:37 crc kubenswrapper[4857]: I1201 21:49:37.418057 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:37 crc kubenswrapper[4857]: I1201 21:49:37.436483 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-wjx7t" podStartSLOduration=8.205684749 podStartE2EDuration="19.436459184s" podCreationTimestamp="2025-12-01 21:49:18 +0000 UTC" firstStartedPulling="2025-12-01 21:49:19.56208763 +0000 UTC m=+878.052149947" lastFinishedPulling="2025-12-01 21:49:30.792862045 +0000 UTC m=+889.282924382" observedRunningTime="2025-12-01 21:49:35.664432314 +0000 UTC m=+894.154494651" watchObservedRunningTime="2025-12-01 21:49:37.436459184 +0000 UTC m=+895.926521501" Dec 01 21:49:39 crc kubenswrapper[4857]: I1201 21:49:39.312381 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:39 crc kubenswrapper[4857]: I1201 21:49:39.362559 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:39 crc kubenswrapper[4857]: I1201 21:49:39.399181 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-2gzlv" Dec 01 21:49:40 crc kubenswrapper[4857]: I1201 21:49:40.874533 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-r64qq" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.218941 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-bbcr5"] Dec 01 21:49:44 crc kubenswrapper[4857]: E1201 21:49:44.219989 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e0d3a3-89e2-41bb-b74d-f0a49b795b72" containerName="extract-content" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.220009 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e0d3a3-89e2-41bb-b74d-f0a49b795b72" containerName="extract-content" Dec 01 21:49:44 crc kubenswrapper[4857]: E1201 21:49:44.220030 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e0d3a3-89e2-41bb-b74d-f0a49b795b72" containerName="registry-server" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.220038 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e0d3a3-89e2-41bb-b74d-f0a49b795b72" containerName="registry-server" Dec 01 21:49:44 crc kubenswrapper[4857]: E1201 21:49:44.220080 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e0d3a3-89e2-41bb-b74d-f0a49b795b72" containerName="extract-utilities" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.220089 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e0d3a3-89e2-41bb-b74d-f0a49b795b72" containerName="extract-utilities" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.220244 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="75e0d3a3-89e2-41bb-b74d-f0a49b795b72" containerName="registry-server" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.220728 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bbcr5" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.222936 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.223519 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.223816 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-4th7n" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.241824 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-bbcr5"] Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.401006 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7spng\" (UniqueName: \"kubernetes.io/projected/89ac3008-d97a-4e78-aff5-e1fc550723bb-kube-api-access-7spng\") pod \"openstack-operator-index-bbcr5\" (UID: \"89ac3008-d97a-4e78-aff5-e1fc550723bb\") " pod="openstack-operators/openstack-operator-index-bbcr5" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.502363 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7spng\" (UniqueName: \"kubernetes.io/projected/89ac3008-d97a-4e78-aff5-e1fc550723bb-kube-api-access-7spng\") pod \"openstack-operator-index-bbcr5\" (UID: \"89ac3008-d97a-4e78-aff5-e1fc550723bb\") " pod="openstack-operators/openstack-operator-index-bbcr5" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.520850 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7spng\" (UniqueName: \"kubernetes.io/projected/89ac3008-d97a-4e78-aff5-e1fc550723bb-kube-api-access-7spng\") pod \"openstack-operator-index-bbcr5\" (UID: \"89ac3008-d97a-4e78-aff5-e1fc550723bb\") " pod="openstack-operators/openstack-operator-index-bbcr5" Dec 01 21:49:44 crc kubenswrapper[4857]: I1201 21:49:44.614008 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bbcr5" Dec 01 21:49:45 crc kubenswrapper[4857]: I1201 21:49:45.074349 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-bbcr5"] Dec 01 21:49:45 crc kubenswrapper[4857]: W1201 21:49:45.082848 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89ac3008_d97a_4e78_aff5_e1fc550723bb.slice/crio-138ad12aa3bbefb4d04fc41e74581920130dda9188f6360bce437541d944d932 WatchSource:0}: Error finding container 138ad12aa3bbefb4d04fc41e74581920130dda9188f6360bce437541d944d932: Status 404 returned error can't find the container with id 138ad12aa3bbefb4d04fc41e74581920130dda9188f6360bce437541d944d932 Dec 01 21:49:45 crc kubenswrapper[4857]: I1201 21:49:45.708518 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bbcr5" event={"ID":"89ac3008-d97a-4e78-aff5-e1fc550723bb","Type":"ContainerStarted","Data":"138ad12aa3bbefb4d04fc41e74581920130dda9188f6360bce437541d944d932"} Dec 01 21:49:47 crc kubenswrapper[4857]: I1201 21:49:47.409010 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:49 crc kubenswrapper[4857]: I1201 21:49:49.281911 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-bbcr5"] Dec 01 21:49:49 crc kubenswrapper[4857]: I1201 21:49:49.316636 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-wjx7t" Dec 01 21:49:49 crc kubenswrapper[4857]: I1201 21:49:49.339441 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-x6nqn" Dec 01 21:49:49 crc kubenswrapper[4857]: I1201 21:49:49.896950 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-m6lxm"] Dec 01 21:49:49 crc kubenswrapper[4857]: I1201 21:49:49.898236 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m6lxm" Dec 01 21:49:49 crc kubenswrapper[4857]: I1201 21:49:49.911160 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-m6lxm"] Dec 01 21:49:50 crc kubenswrapper[4857]: I1201 21:49:50.097208 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlhtp\" (UniqueName: \"kubernetes.io/projected/f8b95808-7d34-4097-bac8-f7e081d9c039-kube-api-access-tlhtp\") pod \"openstack-operator-index-m6lxm\" (UID: \"f8b95808-7d34-4097-bac8-f7e081d9c039\") " pod="openstack-operators/openstack-operator-index-m6lxm" Dec 01 21:49:50 crc kubenswrapper[4857]: I1201 21:49:50.199903 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlhtp\" (UniqueName: \"kubernetes.io/projected/f8b95808-7d34-4097-bac8-f7e081d9c039-kube-api-access-tlhtp\") pod \"openstack-operator-index-m6lxm\" (UID: \"f8b95808-7d34-4097-bac8-f7e081d9c039\") " pod="openstack-operators/openstack-operator-index-m6lxm" Dec 01 21:49:50 crc kubenswrapper[4857]: I1201 21:49:50.233939 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlhtp\" (UniqueName: \"kubernetes.io/projected/f8b95808-7d34-4097-bac8-f7e081d9c039-kube-api-access-tlhtp\") pod \"openstack-operator-index-m6lxm\" (UID: \"f8b95808-7d34-4097-bac8-f7e081d9c039\") " pod="openstack-operators/openstack-operator-index-m6lxm" Dec 01 21:49:50 crc kubenswrapper[4857]: I1201 21:49:50.529592 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m6lxm" Dec 01 21:49:51 crc kubenswrapper[4857]: I1201 21:49:51.489948 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ql4mw"] Dec 01 21:49:51 crc kubenswrapper[4857]: I1201 21:49:51.491532 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ql4mw" podUID="f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" containerName="registry-server" containerID="cri-o://f4fb541a41d4163d50a9319cced538a8c3aec50fb1428dd1c3c4c9989246c7ed" gracePeriod=2 Dec 01 21:49:51 crc kubenswrapper[4857]: I1201 21:49:51.752682 4857 generic.go:334] "Generic (PLEG): container finished" podID="f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" containerID="f4fb541a41d4163d50a9319cced538a8c3aec50fb1428dd1c3c4c9989246c7ed" exitCode=0 Dec 01 21:49:51 crc kubenswrapper[4857]: I1201 21:49:51.752998 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ql4mw" event={"ID":"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53","Type":"ContainerDied","Data":"f4fb541a41d4163d50a9319cced538a8c3aec50fb1428dd1c3c4c9989246c7ed"} Dec 01 21:49:51 crc kubenswrapper[4857]: I1201 21:49:51.908782 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:51 crc kubenswrapper[4857]: I1201 21:49:51.937141 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-m6lxm"] Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.030566 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjd2n\" (UniqueName: \"kubernetes.io/projected/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-kube-api-access-cjd2n\") pod \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\" (UID: \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\") " Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.030645 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-catalog-content\") pod \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\" (UID: \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\") " Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.030699 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-utilities\") pod \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\" (UID: \"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53\") " Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.032175 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-utilities" (OuterVolumeSpecName: "utilities") pod "f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" (UID: "f78ae4f8-29a7-4ac1-b90d-cef7983c0a53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.036418 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-kube-api-access-cjd2n" (OuterVolumeSpecName: "kube-api-access-cjd2n") pod "f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" (UID: "f78ae4f8-29a7-4ac1-b90d-cef7983c0a53"). InnerVolumeSpecName "kube-api-access-cjd2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.053030 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" (UID: "f78ae4f8-29a7-4ac1-b90d-cef7983c0a53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.132670 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjd2n\" (UniqueName: \"kubernetes.io/projected/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-kube-api-access-cjd2n\") on node \"crc\" DevicePath \"\"" Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.132711 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.132722 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.772388 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m6lxm" event={"ID":"f8b95808-7d34-4097-bac8-f7e081d9c039","Type":"ContainerStarted","Data":"8b43d67c3d70ce5c39bffacea5f8639386bd70d0e1a0bd6c16dd60e1acbee86e"} Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.772458 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m6lxm" event={"ID":"f8b95808-7d34-4097-bac8-f7e081d9c039","Type":"ContainerStarted","Data":"1e6d4d81641ecb7d029bf3447cf95456f02c0198c8941fe122314cbad6ca7662"} Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.828178 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bbcr5" event={"ID":"89ac3008-d97a-4e78-aff5-e1fc550723bb","Type":"ContainerStarted","Data":"8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617"} Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.828289 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-bbcr5" podUID="89ac3008-d97a-4e78-aff5-e1fc550723bb" containerName="registry-server" containerID="cri-o://8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617" gracePeriod=2 Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.833121 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ql4mw" event={"ID":"f78ae4f8-29a7-4ac1-b90d-cef7983c0a53","Type":"ContainerDied","Data":"3b02b462bb85d9ed51baeeba2c31aec7b22414b1d506bdd0c3245fdce87ffcfa"} Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.833210 4857 scope.go:117] "RemoveContainer" containerID="f4fb541a41d4163d50a9319cced538a8c3aec50fb1428dd1c3c4c9989246c7ed" Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.833237 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ql4mw" Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.852718 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-m6lxm" podStartSLOduration=3.72898314 podStartE2EDuration="3.852686555s" podCreationTimestamp="2025-12-01 21:49:49 +0000 UTC" firstStartedPulling="2025-12-01 21:49:51.946244652 +0000 UTC m=+910.436306959" lastFinishedPulling="2025-12-01 21:49:52.069948057 +0000 UTC m=+910.560010374" observedRunningTime="2025-12-01 21:49:52.846195343 +0000 UTC m=+911.336257660" watchObservedRunningTime="2025-12-01 21:49:52.852686555 +0000 UTC m=+911.342748912" Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.858763 4857 scope.go:117] "RemoveContainer" containerID="797c6aae26d02cfa40a2f5cd1cb6bd3bb06bbb35e097f45e76fc4b03796677fc" Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.884590 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-bbcr5" podStartSLOduration=2.4443595399999998 podStartE2EDuration="8.884559645s" podCreationTimestamp="2025-12-01 21:49:44 +0000 UTC" firstStartedPulling="2025-12-01 21:49:45.085511027 +0000 UTC m=+903.575573334" lastFinishedPulling="2025-12-01 21:49:51.525711112 +0000 UTC m=+910.015773439" observedRunningTime="2025-12-01 21:49:52.883688946 +0000 UTC m=+911.373751273" watchObservedRunningTime="2025-12-01 21:49:52.884559645 +0000 UTC m=+911.374621962" Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.942915 4857 scope.go:117] "RemoveContainer" containerID="fa4e8f43e6c2359377df6a7d9bdb8eaa0cc2a5336d6290bcb66b877f22d654fe" Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.950144 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ql4mw"] Dec 01 21:49:52 crc kubenswrapper[4857]: I1201 21:49:52.957607 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ql4mw"] Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.232763 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bbcr5" Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.350278 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7spng\" (UniqueName: \"kubernetes.io/projected/89ac3008-d97a-4e78-aff5-e1fc550723bb-kube-api-access-7spng\") pod \"89ac3008-d97a-4e78-aff5-e1fc550723bb\" (UID: \"89ac3008-d97a-4e78-aff5-e1fc550723bb\") " Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.358785 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89ac3008-d97a-4e78-aff5-e1fc550723bb-kube-api-access-7spng" (OuterVolumeSpecName: "kube-api-access-7spng") pod "89ac3008-d97a-4e78-aff5-e1fc550723bb" (UID: "89ac3008-d97a-4e78-aff5-e1fc550723bb"). InnerVolumeSpecName "kube-api-access-7spng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.452117 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7spng\" (UniqueName: \"kubernetes.io/projected/89ac3008-d97a-4e78-aff5-e1fc550723bb-kube-api-access-7spng\") on node \"crc\" DevicePath \"\"" Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.843232 4857 generic.go:334] "Generic (PLEG): container finished" podID="89ac3008-d97a-4e78-aff5-e1fc550723bb" containerID="8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617" exitCode=0 Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.843367 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bbcr5" Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.850624 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" path="/var/lib/kubelet/pods/f78ae4f8-29a7-4ac1-b90d-cef7983c0a53/volumes" Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.851559 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bbcr5" event={"ID":"89ac3008-d97a-4e78-aff5-e1fc550723bb","Type":"ContainerDied","Data":"8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617"} Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.851599 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bbcr5" event={"ID":"89ac3008-d97a-4e78-aff5-e1fc550723bb","Type":"ContainerDied","Data":"138ad12aa3bbefb4d04fc41e74581920130dda9188f6360bce437541d944d932"} Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.851628 4857 scope.go:117] "RemoveContainer" containerID="8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617" Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.950027 4857 scope.go:117] "RemoveContainer" containerID="8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617" Dec 01 21:49:53 crc kubenswrapper[4857]: E1201 21:49:53.950633 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617\": container with ID starting with 8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617 not found: ID does not exist" containerID="8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617" Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.950683 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617"} err="failed to get container status \"8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617\": rpc error: code = NotFound desc = could not find container \"8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617\": container with ID starting with 8727f4bd79845c1a81cd6021a0b1b2fbd0ee3f3a576c927b0705123e7fc58617 not found: ID does not exist" Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.958600 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-bbcr5"] Dec 01 21:49:53 crc kubenswrapper[4857]: I1201 21:49:53.963694 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-bbcr5"] Dec 01 21:49:55 crc kubenswrapper[4857]: I1201 21:49:55.847162 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89ac3008-d97a-4e78-aff5-e1fc550723bb" path="/var/lib/kubelet/pods/89ac3008-d97a-4e78-aff5-e1fc550723bb/volumes" Dec 01 21:50:00 crc kubenswrapper[4857]: I1201 21:50:00.530100 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-m6lxm" Dec 01 21:50:00 crc kubenswrapper[4857]: I1201 21:50:00.530436 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-m6lxm" Dec 01 21:50:00 crc kubenswrapper[4857]: I1201 21:50:00.645209 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-m6lxm" Dec 01 21:50:01 crc kubenswrapper[4857]: I1201 21:50:01.005119 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-m6lxm" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.140633 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms"] Dec 01 21:50:07 crc kubenswrapper[4857]: E1201 21:50:07.141654 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" containerName="extract-utilities" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.141679 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" containerName="extract-utilities" Dec 01 21:50:07 crc kubenswrapper[4857]: E1201 21:50:07.141699 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" containerName="registry-server" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.141712 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" containerName="registry-server" Dec 01 21:50:07 crc kubenswrapper[4857]: E1201 21:50:07.141742 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" containerName="extract-content" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.141757 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" containerName="extract-content" Dec 01 21:50:07 crc kubenswrapper[4857]: E1201 21:50:07.141779 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89ac3008-d97a-4e78-aff5-e1fc550723bb" containerName="registry-server" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.141792 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="89ac3008-d97a-4e78-aff5-e1fc550723bb" containerName="registry-server" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.141997 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="89ac3008-d97a-4e78-aff5-e1fc550723bb" containerName="registry-server" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.142015 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f78ae4f8-29a7-4ac1-b90d-cef7983c0a53" containerName="registry-server" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.143647 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.145774 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-vvwsc" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.151440 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms"] Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.253727 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-util\") pod \"20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms\" (UID: \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\") " pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.253815 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw9wp\" (UniqueName: \"kubernetes.io/projected/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-kube-api-access-kw9wp\") pod \"20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms\" (UID: \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\") " pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.253879 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-bundle\") pod \"20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms\" (UID: \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\") " pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.355736 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-bundle\") pod \"20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms\" (UID: \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\") " pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.355888 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-util\") pod \"20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms\" (UID: \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\") " pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.355939 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw9wp\" (UniqueName: \"kubernetes.io/projected/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-kube-api-access-kw9wp\") pod \"20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms\" (UID: \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\") " pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.356593 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-bundle\") pod \"20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms\" (UID: \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\") " pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.356871 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-util\") pod \"20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms\" (UID: \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\") " pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.381423 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw9wp\" (UniqueName: \"kubernetes.io/projected/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-kube-api-access-kw9wp\") pod \"20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms\" (UID: \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\") " pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.481266 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.863907 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms"] Dec 01 21:50:07 crc kubenswrapper[4857]: I1201 21:50:07.966322 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" event={"ID":"8ff7e678-7e85-4ba7-a5cb-d332cb47564a","Type":"ContainerStarted","Data":"14b06bad2fbeabdc58f8d9704ab7f4801960b53529b15a38b5b86a7e4aeb7a95"} Dec 01 21:50:08 crc kubenswrapper[4857]: I1201 21:50:08.976725 4857 generic.go:334] "Generic (PLEG): container finished" podID="8ff7e678-7e85-4ba7-a5cb-d332cb47564a" containerID="ca951debe4d85b3f0e5c299263c6a2322415bdaf7a5d1258a84cb54dc43723be" exitCode=0 Dec 01 21:50:08 crc kubenswrapper[4857]: I1201 21:50:08.976821 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" event={"ID":"8ff7e678-7e85-4ba7-a5cb-d332cb47564a","Type":"ContainerDied","Data":"ca951debe4d85b3f0e5c299263c6a2322415bdaf7a5d1258a84cb54dc43723be"} Dec 01 21:50:09 crc kubenswrapper[4857]: I1201 21:50:09.992424 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" event={"ID":"8ff7e678-7e85-4ba7-a5cb-d332cb47564a","Type":"ContainerStarted","Data":"201eaeabf8544832a110fdc83a5192f9b133358269b85b4b2ad4f6602f89f366"} Dec 01 21:50:11 crc kubenswrapper[4857]: I1201 21:50:11.015106 4857 generic.go:334] "Generic (PLEG): container finished" podID="8ff7e678-7e85-4ba7-a5cb-d332cb47564a" containerID="201eaeabf8544832a110fdc83a5192f9b133358269b85b4b2ad4f6602f89f366" exitCode=0 Dec 01 21:50:11 crc kubenswrapper[4857]: I1201 21:50:11.015221 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" event={"ID":"8ff7e678-7e85-4ba7-a5cb-d332cb47564a","Type":"ContainerDied","Data":"201eaeabf8544832a110fdc83a5192f9b133358269b85b4b2ad4f6602f89f366"} Dec 01 21:50:12 crc kubenswrapper[4857]: I1201 21:50:12.025779 4857 generic.go:334] "Generic (PLEG): container finished" podID="8ff7e678-7e85-4ba7-a5cb-d332cb47564a" containerID="0498a7276a2b3a3ce04a2467931e756b165be5cb44f4617615831d7b7cad2421" exitCode=0 Dec 01 21:50:12 crc kubenswrapper[4857]: I1201 21:50:12.025882 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" event={"ID":"8ff7e678-7e85-4ba7-a5cb-d332cb47564a","Type":"ContainerDied","Data":"0498a7276a2b3a3ce04a2467931e756b165be5cb44f4617615831d7b7cad2421"} Dec 01 21:50:13 crc kubenswrapper[4857]: I1201 21:50:13.341056 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:13 crc kubenswrapper[4857]: I1201 21:50:13.451060 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw9wp\" (UniqueName: \"kubernetes.io/projected/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-kube-api-access-kw9wp\") pod \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\" (UID: \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\") " Dec 01 21:50:13 crc kubenswrapper[4857]: I1201 21:50:13.451152 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-bundle\") pod \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\" (UID: \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\") " Dec 01 21:50:13 crc kubenswrapper[4857]: I1201 21:50:13.451238 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-util\") pod \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\" (UID: \"8ff7e678-7e85-4ba7-a5cb-d332cb47564a\") " Dec 01 21:50:13 crc kubenswrapper[4857]: I1201 21:50:13.452459 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-bundle" (OuterVolumeSpecName: "bundle") pod "8ff7e678-7e85-4ba7-a5cb-d332cb47564a" (UID: "8ff7e678-7e85-4ba7-a5cb-d332cb47564a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:50:13 crc kubenswrapper[4857]: I1201 21:50:13.460277 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-kube-api-access-kw9wp" (OuterVolumeSpecName: "kube-api-access-kw9wp") pod "8ff7e678-7e85-4ba7-a5cb-d332cb47564a" (UID: "8ff7e678-7e85-4ba7-a5cb-d332cb47564a"). InnerVolumeSpecName "kube-api-access-kw9wp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:50:13 crc kubenswrapper[4857]: I1201 21:50:13.552663 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw9wp\" (UniqueName: \"kubernetes.io/projected/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-kube-api-access-kw9wp\") on node \"crc\" DevicePath \"\"" Dec 01 21:50:13 crc kubenswrapper[4857]: I1201 21:50:13.552721 4857 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:50:13 crc kubenswrapper[4857]: I1201 21:50:13.573754 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-util" (OuterVolumeSpecName: "util") pod "8ff7e678-7e85-4ba7-a5cb-d332cb47564a" (UID: "8ff7e678-7e85-4ba7-a5cb-d332cb47564a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:50:13 crc kubenswrapper[4857]: I1201 21:50:13.654231 4857 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ff7e678-7e85-4ba7-a5cb-d332cb47564a-util\") on node \"crc\" DevicePath \"\"" Dec 01 21:50:14 crc kubenswrapper[4857]: I1201 21:50:14.043733 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" event={"ID":"8ff7e678-7e85-4ba7-a5cb-d332cb47564a","Type":"ContainerDied","Data":"14b06bad2fbeabdc58f8d9704ab7f4801960b53529b15a38b5b86a7e4aeb7a95"} Dec 01 21:50:14 crc kubenswrapper[4857]: I1201 21:50:14.043946 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14b06bad2fbeabdc58f8d9704ab7f4801960b53529b15a38b5b86a7e4aeb7a95" Dec 01 21:50:14 crc kubenswrapper[4857]: I1201 21:50:14.043806 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms" Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.371919 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd"] Dec 01 21:50:19 crc kubenswrapper[4857]: E1201 21:50:19.372833 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff7e678-7e85-4ba7-a5cb-d332cb47564a" containerName="pull" Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.372848 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff7e678-7e85-4ba7-a5cb-d332cb47564a" containerName="pull" Dec 01 21:50:19 crc kubenswrapper[4857]: E1201 21:50:19.372860 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff7e678-7e85-4ba7-a5cb-d332cb47564a" containerName="util" Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.372868 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff7e678-7e85-4ba7-a5cb-d332cb47564a" containerName="util" Dec 01 21:50:19 crc kubenswrapper[4857]: E1201 21:50:19.372892 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff7e678-7e85-4ba7-a5cb-d332cb47564a" containerName="extract" Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.372903 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff7e678-7e85-4ba7-a5cb-d332cb47564a" containerName="extract" Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.373083 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ff7e678-7e85-4ba7-a5cb-d332cb47564a" containerName="extract" Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.373579 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd" Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.382655 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-ll5wf" Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.398887 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd"] Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.432430 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfxsb\" (UniqueName: \"kubernetes.io/projected/1e12f76f-2a0a-46d8-aec6-6d9a7cf7e0dc-kube-api-access-mfxsb\") pod \"openstack-operator-controller-operator-9f7fd9cfd-66kxd\" (UID: \"1e12f76f-2a0a-46d8-aec6-6d9a7cf7e0dc\") " pod="openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd" Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.534330 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfxsb\" (UniqueName: \"kubernetes.io/projected/1e12f76f-2a0a-46d8-aec6-6d9a7cf7e0dc-kube-api-access-mfxsb\") pod \"openstack-operator-controller-operator-9f7fd9cfd-66kxd\" (UID: \"1e12f76f-2a0a-46d8-aec6-6d9a7cf7e0dc\") " pod="openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd" Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.562955 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfxsb\" (UniqueName: \"kubernetes.io/projected/1e12f76f-2a0a-46d8-aec6-6d9a7cf7e0dc-kube-api-access-mfxsb\") pod \"openstack-operator-controller-operator-9f7fd9cfd-66kxd\" (UID: \"1e12f76f-2a0a-46d8-aec6-6d9a7cf7e0dc\") " pod="openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd" Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.695980 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd" Dec 01 21:50:19 crc kubenswrapper[4857]: I1201 21:50:19.899150 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd"] Dec 01 21:50:20 crc kubenswrapper[4857]: I1201 21:50:20.085470 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd" event={"ID":"1e12f76f-2a0a-46d8-aec6-6d9a7cf7e0dc","Type":"ContainerStarted","Data":"c90d718d065ef2bbf2eb8f1c39d35aab01903a3b78b3f7f379f33ce882dbe12b"} Dec 01 21:50:21 crc kubenswrapper[4857]: I1201 21:50:21.640104 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wvg2b"] Dec 01 21:50:21 crc kubenswrapper[4857]: I1201 21:50:21.643244 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:21 crc kubenswrapper[4857]: I1201 21:50:21.656896 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wvg2b"] Dec 01 21:50:21 crc kubenswrapper[4857]: I1201 21:50:21.767877 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qp8z\" (UniqueName: \"kubernetes.io/projected/106df5ec-d85f-4730-93fc-9c8b41c1b86e-kube-api-access-4qp8z\") pod \"certified-operators-wvg2b\" (UID: \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\") " pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:21 crc kubenswrapper[4857]: I1201 21:50:21.768224 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/106df5ec-d85f-4730-93fc-9c8b41c1b86e-catalog-content\") pod \"certified-operators-wvg2b\" (UID: \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\") " pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:21 crc kubenswrapper[4857]: I1201 21:50:21.768368 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/106df5ec-d85f-4730-93fc-9c8b41c1b86e-utilities\") pod \"certified-operators-wvg2b\" (UID: \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\") " pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:21 crc kubenswrapper[4857]: I1201 21:50:21.869326 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qp8z\" (UniqueName: \"kubernetes.io/projected/106df5ec-d85f-4730-93fc-9c8b41c1b86e-kube-api-access-4qp8z\") pod \"certified-operators-wvg2b\" (UID: \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\") " pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:21 crc kubenswrapper[4857]: I1201 21:50:21.869395 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/106df5ec-d85f-4730-93fc-9c8b41c1b86e-catalog-content\") pod \"certified-operators-wvg2b\" (UID: \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\") " pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:21 crc kubenswrapper[4857]: I1201 21:50:21.869461 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/106df5ec-d85f-4730-93fc-9c8b41c1b86e-utilities\") pod \"certified-operators-wvg2b\" (UID: \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\") " pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:21 crc kubenswrapper[4857]: I1201 21:50:21.870177 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/106df5ec-d85f-4730-93fc-9c8b41c1b86e-utilities\") pod \"certified-operators-wvg2b\" (UID: \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\") " pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:21 crc kubenswrapper[4857]: I1201 21:50:21.870426 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/106df5ec-d85f-4730-93fc-9c8b41c1b86e-catalog-content\") pod \"certified-operators-wvg2b\" (UID: \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\") " pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:21 crc kubenswrapper[4857]: I1201 21:50:21.890786 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qp8z\" (UniqueName: \"kubernetes.io/projected/106df5ec-d85f-4730-93fc-9c8b41c1b86e-kube-api-access-4qp8z\") pod \"certified-operators-wvg2b\" (UID: \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\") " pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:22 crc kubenswrapper[4857]: I1201 21:50:22.016977 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:22 crc kubenswrapper[4857]: I1201 21:50:22.478429 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wvg2b"] Dec 01 21:50:23 crc kubenswrapper[4857]: I1201 21:50:23.108030 4857 generic.go:334] "Generic (PLEG): container finished" podID="106df5ec-d85f-4730-93fc-9c8b41c1b86e" containerID="a69451f4b3ab8a9f8212fd4c2e8ae1748c19839e441c1658ca7e70bae815be98" exitCode=0 Dec 01 21:50:23 crc kubenswrapper[4857]: I1201 21:50:23.108358 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvg2b" event={"ID":"106df5ec-d85f-4730-93fc-9c8b41c1b86e","Type":"ContainerDied","Data":"a69451f4b3ab8a9f8212fd4c2e8ae1748c19839e441c1658ca7e70bae815be98"} Dec 01 21:50:23 crc kubenswrapper[4857]: I1201 21:50:23.108384 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvg2b" event={"ID":"106df5ec-d85f-4730-93fc-9c8b41c1b86e","Type":"ContainerStarted","Data":"dca46db2cd615bac86c2cbae527d4508c0c177498f5708225a71c1169a70518d"} Dec 01 21:50:28 crc kubenswrapper[4857]: I1201 21:50:28.145367 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvg2b" event={"ID":"106df5ec-d85f-4730-93fc-9c8b41c1b86e","Type":"ContainerStarted","Data":"940e77fd1989e32831648b293cef0040b3d18e3d1f20e45469101a3cd289662d"} Dec 01 21:50:28 crc kubenswrapper[4857]: I1201 21:50:28.147155 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd" event={"ID":"1e12f76f-2a0a-46d8-aec6-6d9a7cf7e0dc","Type":"ContainerStarted","Data":"08fa459beb96b597affc8bedd0830b8daca9852bcdfdc093f07b75e4a5d6ef7a"} Dec 01 21:50:28 crc kubenswrapper[4857]: I1201 21:50:28.147649 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd" Dec 01 21:50:28 crc kubenswrapper[4857]: I1201 21:50:28.222342 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd" podStartSLOduration=1.594710682 podStartE2EDuration="9.22231929s" podCreationTimestamp="2025-12-01 21:50:19 +0000 UTC" firstStartedPulling="2025-12-01 21:50:19.91622078 +0000 UTC m=+938.406283097" lastFinishedPulling="2025-12-01 21:50:27.543829388 +0000 UTC m=+946.033891705" observedRunningTime="2025-12-01 21:50:28.216335558 +0000 UTC m=+946.706397905" watchObservedRunningTime="2025-12-01 21:50:28.22231929 +0000 UTC m=+946.712381617" Dec 01 21:50:29 crc kubenswrapper[4857]: I1201 21:50:29.155633 4857 generic.go:334] "Generic (PLEG): container finished" podID="106df5ec-d85f-4730-93fc-9c8b41c1b86e" containerID="940e77fd1989e32831648b293cef0040b3d18e3d1f20e45469101a3cd289662d" exitCode=0 Dec 01 21:50:29 crc kubenswrapper[4857]: I1201 21:50:29.155740 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvg2b" event={"ID":"106df5ec-d85f-4730-93fc-9c8b41c1b86e","Type":"ContainerDied","Data":"940e77fd1989e32831648b293cef0040b3d18e3d1f20e45469101a3cd289662d"} Dec 01 21:50:30 crc kubenswrapper[4857]: I1201 21:50:30.167413 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvg2b" event={"ID":"106df5ec-d85f-4730-93fc-9c8b41c1b86e","Type":"ContainerStarted","Data":"4b36f1302f840a31a4f2f625a35dfdd61f1ea43063646e029b0140ba6a8cb921"} Dec 01 21:50:30 crc kubenswrapper[4857]: I1201 21:50:30.190429 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wvg2b" podStartSLOduration=2.621227186 podStartE2EDuration="9.190404674s" podCreationTimestamp="2025-12-01 21:50:21 +0000 UTC" firstStartedPulling="2025-12-01 21:50:23.11020907 +0000 UTC m=+941.600271387" lastFinishedPulling="2025-12-01 21:50:29.679386528 +0000 UTC m=+948.169448875" observedRunningTime="2025-12-01 21:50:30.184798251 +0000 UTC m=+948.674860608" watchObservedRunningTime="2025-12-01 21:50:30.190404674 +0000 UTC m=+948.680467021" Dec 01 21:50:32 crc kubenswrapper[4857]: I1201 21:50:32.018163 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:32 crc kubenswrapper[4857]: I1201 21:50:32.019324 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:32 crc kubenswrapper[4857]: I1201 21:50:32.086478 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:39 crc kubenswrapper[4857]: I1201 21:50:39.699565 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-9f7fd9cfd-66kxd" Dec 01 21:50:42 crc kubenswrapper[4857]: I1201 21:50:42.108794 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:44 crc kubenswrapper[4857]: I1201 21:50:44.423825 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wvg2b"] Dec 01 21:50:44 crc kubenswrapper[4857]: I1201 21:50:44.424387 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wvg2b" podUID="106df5ec-d85f-4730-93fc-9c8b41c1b86e" containerName="registry-server" containerID="cri-o://4b36f1302f840a31a4f2f625a35dfdd61f1ea43063646e029b0140ba6a8cb921" gracePeriod=2 Dec 01 21:50:45 crc kubenswrapper[4857]: I1201 21:50:45.313228 4857 generic.go:334] "Generic (PLEG): container finished" podID="106df5ec-d85f-4730-93fc-9c8b41c1b86e" containerID="4b36f1302f840a31a4f2f625a35dfdd61f1ea43063646e029b0140ba6a8cb921" exitCode=0 Dec 01 21:50:45 crc kubenswrapper[4857]: I1201 21:50:45.313290 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvg2b" event={"ID":"106df5ec-d85f-4730-93fc-9c8b41c1b86e","Type":"ContainerDied","Data":"4b36f1302f840a31a4f2f625a35dfdd61f1ea43063646e029b0140ba6a8cb921"} Dec 01 21:50:45 crc kubenswrapper[4857]: I1201 21:50:45.961176 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.057656 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qp8z\" (UniqueName: \"kubernetes.io/projected/106df5ec-d85f-4730-93fc-9c8b41c1b86e-kube-api-access-4qp8z\") pod \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\" (UID: \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\") " Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.057763 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/106df5ec-d85f-4730-93fc-9c8b41c1b86e-catalog-content\") pod \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\" (UID: \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\") " Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.057983 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/106df5ec-d85f-4730-93fc-9c8b41c1b86e-utilities\") pod \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\" (UID: \"106df5ec-d85f-4730-93fc-9c8b41c1b86e\") " Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.059720 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/106df5ec-d85f-4730-93fc-9c8b41c1b86e-utilities" (OuterVolumeSpecName: "utilities") pod "106df5ec-d85f-4730-93fc-9c8b41c1b86e" (UID: "106df5ec-d85f-4730-93fc-9c8b41c1b86e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.064604 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/106df5ec-d85f-4730-93fc-9c8b41c1b86e-kube-api-access-4qp8z" (OuterVolumeSpecName: "kube-api-access-4qp8z") pod "106df5ec-d85f-4730-93fc-9c8b41c1b86e" (UID: "106df5ec-d85f-4730-93fc-9c8b41c1b86e"). InnerVolumeSpecName "kube-api-access-4qp8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.102911 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/106df5ec-d85f-4730-93fc-9c8b41c1b86e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "106df5ec-d85f-4730-93fc-9c8b41c1b86e" (UID: "106df5ec-d85f-4730-93fc-9c8b41c1b86e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.159977 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/106df5ec-d85f-4730-93fc-9c8b41c1b86e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.160010 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qp8z\" (UniqueName: \"kubernetes.io/projected/106df5ec-d85f-4730-93fc-9c8b41c1b86e-kube-api-access-4qp8z\") on node \"crc\" DevicePath \"\"" Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.160023 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/106df5ec-d85f-4730-93fc-9c8b41c1b86e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.323648 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvg2b" event={"ID":"106df5ec-d85f-4730-93fc-9c8b41c1b86e","Type":"ContainerDied","Data":"dca46db2cd615bac86c2cbae527d4508c0c177498f5708225a71c1169a70518d"} Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.323726 4857 scope.go:117] "RemoveContainer" containerID="4b36f1302f840a31a4f2f625a35dfdd61f1ea43063646e029b0140ba6a8cb921" Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.323758 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wvg2b" Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.349485 4857 scope.go:117] "RemoveContainer" containerID="940e77fd1989e32831648b293cef0040b3d18e3d1f20e45469101a3cd289662d" Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.374506 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wvg2b"] Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.379221 4857 scope.go:117] "RemoveContainer" containerID="a69451f4b3ab8a9f8212fd4c2e8ae1748c19839e441c1658ca7e70bae815be98" Dec 01 21:50:46 crc kubenswrapper[4857]: I1201 21:50:46.381059 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wvg2b"] Dec 01 21:50:47 crc kubenswrapper[4857]: I1201 21:50:47.851612 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="106df5ec-d85f-4730-93fc-9c8b41c1b86e" path="/var/lib/kubelet/pods/106df5ec-d85f-4730-93fc-9c8b41c1b86e/volumes" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.170264 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm"] Dec 01 21:51:06 crc kubenswrapper[4857]: E1201 21:51:06.171321 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106df5ec-d85f-4730-93fc-9c8b41c1b86e" containerName="extract-content" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.171336 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="106df5ec-d85f-4730-93fc-9c8b41c1b86e" containerName="extract-content" Dec 01 21:51:06 crc kubenswrapper[4857]: E1201 21:51:06.171359 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106df5ec-d85f-4730-93fc-9c8b41c1b86e" containerName="registry-server" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.171367 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="106df5ec-d85f-4730-93fc-9c8b41c1b86e" containerName="registry-server" Dec 01 21:51:06 crc kubenswrapper[4857]: E1201 21:51:06.171386 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106df5ec-d85f-4730-93fc-9c8b41c1b86e" containerName="extract-utilities" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.171396 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="106df5ec-d85f-4730-93fc-9c8b41c1b86e" containerName="extract-utilities" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.171534 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="106df5ec-d85f-4730-93fc-9c8b41c1b86e" containerName="registry-server" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.172317 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.174614 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.175129 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-ggdwd" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.175597 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.177907 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-jnslx" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.192018 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.209391 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.217047 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.218169 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.219636 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-pbrwn" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.221337 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.252035 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.269843 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sbk7\" (UniqueName: \"kubernetes.io/projected/d3a93b19-a249-499a-89fc-9b2a1a69c1e0-kube-api-access-5sbk7\") pod \"cinder-operator-controller-manager-859b6ccc6-5gqcm\" (UID: \"d3a93b19-a249-499a-89fc-9b2a1a69c1e0\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.313792 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-2pbjm" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.323811 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.327000 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.329291 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.333444 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-4md6s" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.333601 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.338858 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.340224 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.355404 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-54xtc"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.359622 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-7452p" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.363732 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.363949 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.367388 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-wwbxg" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.368011 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.380835 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ptdl\" (UniqueName: \"kubernetes.io/projected/df91f733-711d-43b0-98a0-b3fa6b119fbe-kube-api-access-5ptdl\") pod \"glance-operator-controller-manager-668d9c48b9-9qwbq\" (UID: \"df91f733-711d-43b0-98a0-b3fa6b119fbe\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.380884 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mdll\" (UniqueName: \"kubernetes.io/projected/393fd1a5-e3b9-44d6-a9ef-1447a0f9f465-kube-api-access-2mdll\") pod \"designate-operator-controller-manager-78b4bc895b-dz5k5\" (UID: \"393fd1a5-e3b9-44d6-a9ef-1447a0f9f465\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.380909 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sbk7\" (UniqueName: \"kubernetes.io/projected/d3a93b19-a249-499a-89fc-9b2a1a69c1e0-kube-api-access-5sbk7\") pod \"cinder-operator-controller-manager-859b6ccc6-5gqcm\" (UID: \"d3a93b19-a249-499a-89fc-9b2a1a69c1e0\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.380961 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvgg5\" (UniqueName: \"kubernetes.io/projected/80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4-kube-api-access-wvgg5\") pod \"barbican-operator-controller-manager-7d9dfd778-rqfzk\" (UID: \"80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.383152 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.430123 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-54xtc"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.443095 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sbk7\" (UniqueName: \"kubernetes.io/projected/d3a93b19-a249-499a-89fc-9b2a1a69c1e0-kube-api-access-5sbk7\") pod \"cinder-operator-controller-manager-859b6ccc6-5gqcm\" (UID: \"d3a93b19-a249-499a-89fc-9b2a1a69c1e0\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.456122 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.457157 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.459567 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.461217 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.474513 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-9v5fw" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.474980 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-jlkgd" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.485371 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.486775 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ptdl\" (UniqueName: \"kubernetes.io/projected/df91f733-711d-43b0-98a0-b3fa6b119fbe-kube-api-access-5ptdl\") pod \"glance-operator-controller-manager-668d9c48b9-9qwbq\" (UID: \"df91f733-711d-43b0-98a0-b3fa6b119fbe\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.486823 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n69m6\" (UniqueName: \"kubernetes.io/projected/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-kube-api-access-n69m6\") pod \"infra-operator-controller-manager-57548d458d-54xtc\" (UID: \"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.486858 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mdll\" (UniqueName: \"kubernetes.io/projected/393fd1a5-e3b9-44d6-a9ef-1447a0f9f465-kube-api-access-2mdll\") pod \"designate-operator-controller-manager-78b4bc895b-dz5k5\" (UID: \"393fd1a5-e3b9-44d6-a9ef-1447a0f9f465\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.486905 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert\") pod \"infra-operator-controller-manager-57548d458d-54xtc\" (UID: \"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.486927 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp7nf\" (UniqueName: \"kubernetes.io/projected/068a21df-001f-45e6-9946-68f67bad7d43-kube-api-access-mp7nf\") pod \"heat-operator-controller-manager-5f64f6f8bb-zkqw9\" (UID: \"068a21df-001f-45e6-9946-68f67bad7d43\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.486995 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvgg5\" (UniqueName: \"kubernetes.io/projected/80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4-kube-api-access-wvgg5\") pod \"barbican-operator-controller-manager-7d9dfd778-rqfzk\" (UID: \"80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.487366 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwkhk\" (UniqueName: \"kubernetes.io/projected/b972200d-235d-4ae5-a71d-dd206e4f6025-kube-api-access-nwkhk\") pod \"horizon-operator-controller-manager-68c6d99b8f-4g68b\" (UID: \"b972200d-235d-4ae5-a71d-dd206e4f6025\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.502110 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.502407 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.537146 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.538180 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.541029 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-pvdqc" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.553105 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvgg5\" (UniqueName: \"kubernetes.io/projected/80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4-kube-api-access-wvgg5\") pod \"barbican-operator-controller-manager-7d9dfd778-rqfzk\" (UID: \"80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.558673 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ptdl\" (UniqueName: \"kubernetes.io/projected/df91f733-711d-43b0-98a0-b3fa6b119fbe-kube-api-access-5ptdl\") pod \"glance-operator-controller-manager-668d9c48b9-9qwbq\" (UID: \"df91f733-711d-43b0-98a0-b3fa6b119fbe\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.579807 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.580885 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.581802 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mdll\" (UniqueName: \"kubernetes.io/projected/393fd1a5-e3b9-44d6-a9ef-1447a0f9f465-kube-api-access-2mdll\") pod \"designate-operator-controller-manager-78b4bc895b-dz5k5\" (UID: \"393fd1a5-e3b9-44d6-a9ef-1447a0f9f465\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.592472 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-bpjq5" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.593260 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqmqx\" (UniqueName: \"kubernetes.io/projected/dcef9388-5d9b-4172-bb80-ec69b3e98dd0-kube-api-access-lqmqx\") pod \"keystone-operator-controller-manager-546d4bdf48-vtcx7\" (UID: \"dcef9388-5d9b-4172-bb80-ec69b3e98dd0\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.593325 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwkhk\" (UniqueName: \"kubernetes.io/projected/b972200d-235d-4ae5-a71d-dd206e4f6025-kube-api-access-nwkhk\") pod \"horizon-operator-controller-manager-68c6d99b8f-4g68b\" (UID: \"b972200d-235d-4ae5-a71d-dd206e4f6025\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.593353 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7mgw\" (UniqueName: \"kubernetes.io/projected/5dd6bfb5-2edc-49b4-8877-85ab5722cdf0-kube-api-access-m7mgw\") pod \"ironic-operator-controller-manager-6c548fd776-j277f\" (UID: \"5dd6bfb5-2edc-49b4-8877-85ab5722cdf0\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.593374 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n69m6\" (UniqueName: \"kubernetes.io/projected/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-kube-api-access-n69m6\") pod \"infra-operator-controller-manager-57548d458d-54xtc\" (UID: \"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.593404 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert\") pod \"infra-operator-controller-manager-57548d458d-54xtc\" (UID: \"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.593422 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp7nf\" (UniqueName: \"kubernetes.io/projected/068a21df-001f-45e6-9946-68f67bad7d43-kube-api-access-mp7nf\") pod \"heat-operator-controller-manager-5f64f6f8bb-zkqw9\" (UID: \"068a21df-001f-45e6-9946-68f67bad7d43\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9" Dec 01 21:51:06 crc kubenswrapper[4857]: E1201 21:51:06.593887 4857 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 21:51:06 crc kubenswrapper[4857]: E1201 21:51:06.593927 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert podName:23e7348c-65d9-4c4b-9c1e-d86cc3b55bff nodeName:}" failed. No retries permitted until 2025-12-01 21:51:07.093909522 +0000 UTC m=+985.583971839 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert") pod "infra-operator-controller-manager-57548d458d-54xtc" (UID: "23e7348c-65d9-4c4b-9c1e-d86cc3b55bff") : secret "infra-operator-webhook-server-cert" not found Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.626831 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.629818 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.631539 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwkhk\" (UniqueName: \"kubernetes.io/projected/b972200d-235d-4ae5-a71d-dd206e4f6025-kube-api-access-nwkhk\") pod \"horizon-operator-controller-manager-68c6d99b8f-4g68b\" (UID: \"b972200d-235d-4ae5-a71d-dd206e4f6025\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.658516 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.683434 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.684340 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.684559 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.695980 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp7nf\" (UniqueName: \"kubernetes.io/projected/068a21df-001f-45e6-9946-68f67bad7d43-kube-api-access-mp7nf\") pod \"heat-operator-controller-manager-5f64f6f8bb-zkqw9\" (UID: \"068a21df-001f-45e6-9946-68f67bad7d43\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.696102 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.697197 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7mgw\" (UniqueName: \"kubernetes.io/projected/5dd6bfb5-2edc-49b4-8877-85ab5722cdf0-kube-api-access-m7mgw\") pod \"ironic-operator-controller-manager-6c548fd776-j277f\" (UID: \"5dd6bfb5-2edc-49b4-8877-85ab5722cdf0\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.697288 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62vjt\" (UniqueName: \"kubernetes.io/projected/b13a13cc-0689-480a-bc95-e19aba327f9c-kube-api-access-62vjt\") pod \"mariadb-operator-controller-manager-56bbcc9d85-f9d2t\" (UID: \"b13a13cc-0689-480a-bc95-e19aba327f9c\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.697381 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqmqx\" (UniqueName: \"kubernetes.io/projected/dcef9388-5d9b-4172-bb80-ec69b3e98dd0-kube-api-access-lqmqx\") pod \"keystone-operator-controller-manager-546d4bdf48-vtcx7\" (UID: \"dcef9388-5d9b-4172-bb80-ec69b3e98dd0\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.697447 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6f7l\" (UniqueName: \"kubernetes.io/projected/92c43db5-8128-4269-9ad4-a263de88f450-kube-api-access-x6f7l\") pod \"manila-operator-controller-manager-6546668bfd-b7ksb\" (UID: \"92c43db5-8128-4269-9ad4-a263de88f450\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.729560 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-cncd6"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.749656 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n69m6\" (UniqueName: \"kubernetes.io/projected/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-kube-api-access-n69m6\") pod \"infra-operator-controller-manager-57548d458d-54xtc\" (UID: \"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.751449 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-pnb5b" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.765627 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.766626 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cncd6" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.766649 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.772835 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7mgw\" (UniqueName: \"kubernetes.io/projected/5dd6bfb5-2edc-49b4-8877-85ab5722cdf0-kube-api-access-m7mgw\") pod \"ironic-operator-controller-manager-6c548fd776-j277f\" (UID: \"5dd6bfb5-2edc-49b4-8877-85ab5722cdf0\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.773431 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-q5vln" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.775741 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-px4bw" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.796772 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqmqx\" (UniqueName: \"kubernetes.io/projected/dcef9388-5d9b-4172-bb80-ec69b3e98dd0-kube-api-access-lqmqx\") pod \"keystone-operator-controller-manager-546d4bdf48-vtcx7\" (UID: \"dcef9388-5d9b-4172-bb80-ec69b3e98dd0\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.804530 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62vjt\" (UniqueName: \"kubernetes.io/projected/b13a13cc-0689-480a-bc95-e19aba327f9c-kube-api-access-62vjt\") pod \"mariadb-operator-controller-manager-56bbcc9d85-f9d2t\" (UID: \"b13a13cc-0689-480a-bc95-e19aba327f9c\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.810731 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-557w7\" (UniqueName: \"kubernetes.io/projected/188a7d67-ffec-4608-9b5f-cc33972e6fed-kube-api-access-557w7\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-p6v4v\" (UID: \"188a7d67-ffec-4608-9b5f-cc33972e6fed\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.811014 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6f7l\" (UniqueName: \"kubernetes.io/projected/92c43db5-8128-4269-9ad4-a263de88f450-kube-api-access-x6f7l\") pod \"manila-operator-controller-manager-6546668bfd-b7ksb\" (UID: \"92c43db5-8128-4269-9ad4-a263de88f450\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.809066 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.828914 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.837805 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.840942 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-cncd6"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.853497 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.862646 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62vjt\" (UniqueName: \"kubernetes.io/projected/b13a13cc-0689-480a-bc95-e19aba327f9c-kube-api-access-62vjt\") pod \"mariadb-operator-controller-manager-56bbcc9d85-f9d2t\" (UID: \"b13a13cc-0689-480a-bc95-e19aba327f9c\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.863989 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6f7l\" (UniqueName: \"kubernetes.io/projected/92c43db5-8128-4269-9ad4-a263de88f450-kube-api-access-x6f7l\") pod \"manila-operator-controller-manager-6546668bfd-b7ksb\" (UID: \"92c43db5-8128-4269-9ad4-a263de88f450\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.868787 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.871100 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.872545 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-v29rp" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.883191 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.887696 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.898937 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.902551 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-mww86"] Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.945659 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-p6lj4" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.958299 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" Dec 01 21:51:06 crc kubenswrapper[4857]: I1201 21:51:06.986728 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-vzmm2" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.001171 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phmqv\" (UniqueName: \"kubernetes.io/projected/246a4bf0-515b-46f8-bd7f-89e7e961d774-kube-api-access-phmqv\") pod \"octavia-operator-controller-manager-998648c74-cncd6\" (UID: \"246a4bf0-515b-46f8-bd7f-89e7e961d774\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-cncd6" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.001305 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28nhf\" (UniqueName: \"kubernetes.io/projected/de53c246-6e20-4b67-be68-1c5356fd46d4-kube-api-access-28nhf\") pod \"nova-operator-controller-manager-697bc559fc-vbvnt\" (UID: \"de53c246-6e20-4b67-be68-1c5356fd46d4\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.001488 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-557w7\" (UniqueName: \"kubernetes.io/projected/188a7d67-ffec-4608-9b5f-cc33972e6fed-kube-api-access-557w7\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-p6v4v\" (UID: \"188a7d67-ffec-4608-9b5f-cc33972e6fed\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.004828 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.085697 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.093728 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.106476 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vwnb\" (UniqueName: \"kubernetes.io/projected/c5ab847c-10d3-482d-8c70-e3ab439e7b5d-kube-api-access-9vwnb\") pod \"placement-operator-controller-manager-78f8948974-mww86\" (UID: \"c5ab847c-10d3-482d-8c70-e3ab439e7b5d\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.106548 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert\") pod \"infra-operator-controller-manager-57548d458d-54xtc\" (UID: \"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.106575 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28nhf\" (UniqueName: \"kubernetes.io/projected/de53c246-6e20-4b67-be68-1c5356fd46d4-kube-api-access-28nhf\") pod \"nova-operator-controller-manager-697bc559fc-vbvnt\" (UID: \"de53c246-6e20-4b67-be68-1c5356fd46d4\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.106619 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c\" (UID: \"1c2d54cf-e53d-483d-9b00-6f03858806ab\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.106683 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmfpc\" (UniqueName: \"kubernetes.io/projected/f6e43ef6-1269-40f5-af4d-fe2f30dea319-kube-api-access-gmfpc\") pod \"ovn-operator-controller-manager-b6456fdb6-7vmng\" (UID: \"f6e43ef6-1269-40f5-af4d-fe2f30dea319\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng" Dec 01 21:51:07 crc kubenswrapper[4857]: E1201 21:51:07.106730 4857 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 21:51:07 crc kubenswrapper[4857]: E1201 21:51:07.106818 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert podName:23e7348c-65d9-4c4b-9c1e-d86cc3b55bff nodeName:}" failed. No retries permitted until 2025-12-01 21:51:08.106768261 +0000 UTC m=+986.596830568 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert") pod "infra-operator-controller-manager-57548d458d-54xtc" (UID: "23e7348c-65d9-4c4b-9c1e-d86cc3b55bff") : secret "infra-operator-webhook-server-cert" not found Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.106742 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phmqv\" (UniqueName: \"kubernetes.io/projected/246a4bf0-515b-46f8-bd7f-89e7e961d774-kube-api-access-phmqv\") pod \"octavia-operator-controller-manager-998648c74-cncd6\" (UID: \"246a4bf0-515b-46f8-bd7f-89e7e961d774\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-cncd6" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.107111 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68vps\" (UniqueName: \"kubernetes.io/projected/1c2d54cf-e53d-483d-9b00-6f03858806ab-kube-api-access-68vps\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c\" (UID: \"1c2d54cf-e53d-483d-9b00-6f03858806ab\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.119374 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.149800 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.150299 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.153019 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-mww86"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.164376 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.165464 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.166800 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.167466 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.180380 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.184381 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-c82xc" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.185724 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phmqv\" (UniqueName: \"kubernetes.io/projected/246a4bf0-515b-46f8-bd7f-89e7e961d774-kube-api-access-phmqv\") pod \"octavia-operator-controller-manager-998648c74-cncd6\" (UID: \"246a4bf0-515b-46f8-bd7f-89e7e961d774\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-cncd6" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.185971 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-5vg6v" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.207261 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.207604 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cncd6" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.212019 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28nhf\" (UniqueName: \"kubernetes.io/projected/de53c246-6e20-4b67-be68-1c5356fd46d4-kube-api-access-28nhf\") pod \"nova-operator-controller-manager-697bc559fc-vbvnt\" (UID: \"de53c246-6e20-4b67-be68-1c5356fd46d4\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.213925 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c\" (UID: \"1c2d54cf-e53d-483d-9b00-6f03858806ab\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.214019 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmfpc\" (UniqueName: \"kubernetes.io/projected/f6e43ef6-1269-40f5-af4d-fe2f30dea319-kube-api-access-gmfpc\") pod \"ovn-operator-controller-manager-b6456fdb6-7vmng\" (UID: \"f6e43ef6-1269-40f5-af4d-fe2f30dea319\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.214133 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68vps\" (UniqueName: \"kubernetes.io/projected/1c2d54cf-e53d-483d-9b00-6f03858806ab-kube-api-access-68vps\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c\" (UID: \"1c2d54cf-e53d-483d-9b00-6f03858806ab\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.214165 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47g5g\" (UniqueName: \"kubernetes.io/projected/cdd278ff-9eb6-45f7-977d-02944f6f312a-kube-api-access-47g5g\") pod \"swift-operator-controller-manager-5f8c65bbfc-wjnzv\" (UID: \"cdd278ff-9eb6-45f7-977d-02944f6f312a\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.214208 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vwnb\" (UniqueName: \"kubernetes.io/projected/c5ab847c-10d3-482d-8c70-e3ab439e7b5d-kube-api-access-9vwnb\") pod \"placement-operator-controller-manager-78f8948974-mww86\" (UID: \"c5ab847c-10d3-482d-8c70-e3ab439e7b5d\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.214267 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjlql\" (UniqueName: \"kubernetes.io/projected/ff230e58-1711-4638-89d6-daf91fb3bfc1-kube-api-access-xjlql\") pod \"telemetry-operator-controller-manager-7f765fc7d6-x797d\" (UID: \"ff230e58-1711-4638-89d6-daf91fb3bfc1\") " pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" Dec 01 21:51:07 crc kubenswrapper[4857]: E1201 21:51:07.214415 4857 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 21:51:07 crc kubenswrapper[4857]: E1201 21:51:07.214458 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert podName:1c2d54cf-e53d-483d-9b00-6f03858806ab nodeName:}" failed. No retries permitted until 2025-12-01 21:51:07.714442808 +0000 UTC m=+986.204505125 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" (UID: "1c2d54cf-e53d-483d-9b00-6f03858806ab") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.221877 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-557w7\" (UniqueName: \"kubernetes.io/projected/188a7d67-ffec-4608-9b5f-cc33972e6fed-kube-api-access-557w7\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-p6v4v\" (UID: \"188a7d67-ffec-4608-9b5f-cc33972e6fed\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.224106 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wccvn"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.232931 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wccvn" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.235471 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-zgk2x" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.244592 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wccvn"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.268664 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vwnb\" (UniqueName: \"kubernetes.io/projected/c5ab847c-10d3-482d-8c70-e3ab439e7b5d-kube-api-access-9vwnb\") pod \"placement-operator-controller-manager-78f8948974-mww86\" (UID: \"c5ab847c-10d3-482d-8c70-e3ab439e7b5d\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.268744 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.271457 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68vps\" (UniqueName: \"kubernetes.io/projected/1c2d54cf-e53d-483d-9b00-6f03858806ab-kube-api-access-68vps\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c\" (UID: \"1c2d54cf-e53d-483d-9b00-6f03858806ab\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.272003 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.276411 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmfpc\" (UniqueName: \"kubernetes.io/projected/f6e43ef6-1269-40f5-af4d-fe2f30dea319-kube-api-access-gmfpc\") pod \"ovn-operator-controller-manager-b6456fdb6-7vmng\" (UID: \"f6e43ef6-1269-40f5-af4d-fe2f30dea319\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.281370 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.285809 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-cvhnr" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.326823 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.328336 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47g5g\" (UniqueName: \"kubernetes.io/projected/cdd278ff-9eb6-45f7-977d-02944f6f312a-kube-api-access-47g5g\") pod \"swift-operator-controller-manager-5f8c65bbfc-wjnzv\" (UID: \"cdd278ff-9eb6-45f7-977d-02944f6f312a\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.328377 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hp7q\" (UniqueName: \"kubernetes.io/projected/4c8d087e-e00c-4dad-87c7-06ab1017abaf-kube-api-access-6hp7q\") pod \"test-operator-controller-manager-5854674fcc-wccvn\" (UID: \"4c8d087e-e00c-4dad-87c7-06ab1017abaf\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wccvn" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.328431 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjlql\" (UniqueName: \"kubernetes.io/projected/ff230e58-1711-4638-89d6-daf91fb3bfc1-kube-api-access-xjlql\") pod \"telemetry-operator-controller-manager-7f765fc7d6-x797d\" (UID: \"ff230e58-1711-4638-89d6-daf91fb3bfc1\") " pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.352183 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjlql\" (UniqueName: \"kubernetes.io/projected/ff230e58-1711-4638-89d6-daf91fb3bfc1-kube-api-access-xjlql\") pod \"telemetry-operator-controller-manager-7f765fc7d6-x797d\" (UID: \"ff230e58-1711-4638-89d6-daf91fb3bfc1\") " pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.353699 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47g5g\" (UniqueName: \"kubernetes.io/projected/cdd278ff-9eb6-45f7-977d-02944f6f312a-kube-api-access-47g5g\") pod \"swift-operator-controller-manager-5f8c65bbfc-wjnzv\" (UID: \"cdd278ff-9eb6-45f7-977d-02944f6f312a\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.401012 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.401967 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.403498 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.406926 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.409842 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.409891 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-bcgs6" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.410035 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.421664 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.430222 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hp7q\" (UniqueName: \"kubernetes.io/projected/4c8d087e-e00c-4dad-87c7-06ab1017abaf-kube-api-access-6hp7q\") pod \"test-operator-controller-manager-5854674fcc-wccvn\" (UID: \"4c8d087e-e00c-4dad-87c7-06ab1017abaf\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wccvn" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.430352 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96xzx\" (UniqueName: \"kubernetes.io/projected/8a6276de-e994-4f62-a625-906c9a5f4f5d-kube-api-access-96xzx\") pod \"watcher-operator-controller-manager-769dc69bc-vh887\" (UID: \"8a6276de-e994-4f62-a625-906c9a5f4f5d\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.433557 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.434625 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.442838 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-jcg2w" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.444977 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh"] Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.449650 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.523146 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.531783 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v95q5\" (UniqueName: \"kubernetes.io/projected/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-kube-api-access-v95q5\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.531839 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.531941 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndrc2\" (UniqueName: \"kubernetes.io/projected/1d490d1b-558b-49fb-9b0f-65d0f82b0148-kube-api-access-ndrc2\") pod \"rabbitmq-cluster-operator-manager-668c99d594-gdqjh\" (UID: \"1d490d1b-558b-49fb-9b0f-65d0f82b0148\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.532058 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96xzx\" (UniqueName: \"kubernetes.io/projected/8a6276de-e994-4f62-a625-906c9a5f4f5d-kube-api-access-96xzx\") pod \"watcher-operator-controller-manager-769dc69bc-vh887\" (UID: \"8a6276de-e994-4f62-a625-906c9a5f4f5d\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.532112 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.556370 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hp7q\" (UniqueName: \"kubernetes.io/projected/4c8d087e-e00c-4dad-87c7-06ab1017abaf-kube-api-access-6hp7q\") pod \"test-operator-controller-manager-5854674fcc-wccvn\" (UID: \"4c8d087e-e00c-4dad-87c7-06ab1017abaf\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wccvn" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.567240 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96xzx\" (UniqueName: \"kubernetes.io/projected/8a6276de-e994-4f62-a625-906c9a5f4f5d-kube-api-access-96xzx\") pod \"watcher-operator-controller-manager-769dc69bc-vh887\" (UID: \"8a6276de-e994-4f62-a625-906c9a5f4f5d\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.569947 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.588695 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wccvn" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.632955 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.633019 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndrc2\" (UniqueName: \"kubernetes.io/projected/1d490d1b-558b-49fb-9b0f-65d0f82b0148-kube-api-access-ndrc2\") pod \"rabbitmq-cluster-operator-manager-668c99d594-gdqjh\" (UID: \"1d490d1b-558b-49fb-9b0f-65d0f82b0148\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.633145 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:07 crc kubenswrapper[4857]: E1201 21:51:07.633180 4857 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 21:51:07 crc kubenswrapper[4857]: E1201 21:51:07.633262 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs podName:0a4c6432-bb6c-41b6-9981-ef6ac1354ae3 nodeName:}" failed. No retries permitted until 2025-12-01 21:51:08.133237824 +0000 UTC m=+986.623300231 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs") pod "openstack-operator-controller-manager-777b97d487-s7g6x" (UID: "0a4c6432-bb6c-41b6-9981-ef6ac1354ae3") : secret "webhook-server-cert" not found Dec 01 21:51:07 crc kubenswrapper[4857]: E1201 21:51:07.633286 4857 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 21:51:07 crc kubenswrapper[4857]: E1201 21:51:07.633348 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs podName:0a4c6432-bb6c-41b6-9981-ef6ac1354ae3 nodeName:}" failed. No retries permitted until 2025-12-01 21:51:08.133326496 +0000 UTC m=+986.623388903 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs") pod "openstack-operator-controller-manager-777b97d487-s7g6x" (UID: "0a4c6432-bb6c-41b6-9981-ef6ac1354ae3") : secret "metrics-server-cert" not found Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.633402 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v95q5\" (UniqueName: \"kubernetes.io/projected/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-kube-api-access-v95q5\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.643488 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.659205 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v95q5\" (UniqueName: \"kubernetes.io/projected/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-kube-api-access-v95q5\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.724351 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndrc2\" (UniqueName: \"kubernetes.io/projected/1d490d1b-558b-49fb-9b0f-65d0f82b0148-kube-api-access-ndrc2\") pod \"rabbitmq-cluster-operator-manager-668c99d594-gdqjh\" (UID: \"1d490d1b-558b-49fb-9b0f-65d0f82b0148\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.734679 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c\" (UID: \"1c2d54cf-e53d-483d-9b00-6f03858806ab\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:07 crc kubenswrapper[4857]: E1201 21:51:07.734845 4857 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 21:51:07 crc kubenswrapper[4857]: E1201 21:51:07.734920 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert podName:1c2d54cf-e53d-483d-9b00-6f03858806ab nodeName:}" failed. No retries permitted until 2025-12-01 21:51:08.734902193 +0000 UTC m=+987.224964510 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" (UID: "1c2d54cf-e53d-483d-9b00-6f03858806ab") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.791964 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh" Dec 01 21:51:07 crc kubenswrapper[4857]: I1201 21:51:07.866755 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm"] Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.154548 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.155001 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert\") pod \"infra-operator-controller-manager-57548d458d-54xtc\" (UID: \"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.155097 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.155280 4857 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.155356 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs podName:0a4c6432-bb6c-41b6-9981-ef6ac1354ae3 nodeName:}" failed. No retries permitted until 2025-12-01 21:51:09.155334829 +0000 UTC m=+987.645397146 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs") pod "openstack-operator-controller-manager-777b97d487-s7g6x" (UID: "0a4c6432-bb6c-41b6-9981-ef6ac1354ae3") : secret "metrics-server-cert" not found Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.155848 4857 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.155894 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs podName:0a4c6432-bb6c-41b6-9981-ef6ac1354ae3 nodeName:}" failed. No retries permitted until 2025-12-01 21:51:09.155883942 +0000 UTC m=+987.645946269 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs") pod "openstack-operator-controller-manager-777b97d487-s7g6x" (UID: "0a4c6432-bb6c-41b6-9981-ef6ac1354ae3") : secret "webhook-server-cert" not found Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.155948 4857 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.155976 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert podName:23e7348c-65d9-4c4b-9c1e-d86cc3b55bff nodeName:}" failed. No retries permitted until 2025-12-01 21:51:10.155967324 +0000 UTC m=+988.646029641 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert") pod "infra-operator-controller-manager-57548d458d-54xtc" (UID: "23e7348c-65d9-4c4b-9c1e-d86cc3b55bff") : secret "infra-operator-webhook-server-cert" not found Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.277954 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk"] Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.284837 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b"] Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.310406 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq"] Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.409345 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5"] Dec 01 21:51:08 crc kubenswrapper[4857]: W1201 21:51:08.412602 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod393fd1a5_e3b9_44d6_a9ef_1447a0f9f465.slice/crio-071bf6d68093330afc592cf2ebb2287d948b58c55c56ff5267164308e785e367 WatchSource:0}: Error finding container 071bf6d68093330afc592cf2ebb2287d948b58c55c56ff5267164308e785e367: Status 404 returned error can't find the container with id 071bf6d68093330afc592cf2ebb2287d948b58c55c56ff5267164308e785e367 Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.452376 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f"] Dec 01 21:51:08 crc kubenswrapper[4857]: W1201 21:51:08.480614 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5dd6bfb5_2edc_49b4_8877_85ab5722cdf0.slice/crio-a2781914a757c1b6efb9d567978e48fd0f730a91afed5b58bd3f32a274a0d319 WatchSource:0}: Error finding container a2781914a757c1b6efb9d567978e48fd0f730a91afed5b58bd3f32a274a0d319: Status 404 returned error can't find the container with id a2781914a757c1b6efb9d567978e48fd0f730a91afed5b58bd3f32a274a0d319 Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.577735 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng"] Dec 01 21:51:08 crc kubenswrapper[4857]: W1201 21:51:08.587221 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6e43ef6_1269_40f5_af4d_fe2f30dea319.slice/crio-f3b854e1b0db0de2a2989cbf10130964334c4ecf292a2741292d54e6034b0d6c WatchSource:0}: Error finding container f3b854e1b0db0de2a2989cbf10130964334c4ecf292a2741292d54e6034b0d6c: Status 404 returned error can't find the container with id f3b854e1b0db0de2a2989cbf10130964334c4ecf292a2741292d54e6034b0d6c Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.645190 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb"] Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.657153 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t"] Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.667304 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt"] Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.676540 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t" event={"ID":"b13a13cc-0689-480a-bc95-e19aba327f9c","Type":"ContainerStarted","Data":"0c5d96793b3d6ebfacb27a8099168eaf576ecf51121a8ba71d06e6e82b7dbf12"} Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.683798 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk" event={"ID":"80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4","Type":"ContainerStarted","Data":"4687192115a80bddddb0b4fd259a89898fbe3e6b52560494aa9955f6a8d26e05"} Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.685460 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5" event={"ID":"393fd1a5-e3b9-44d6-a9ef-1447a0f9f465","Type":"ContainerStarted","Data":"071bf6d68093330afc592cf2ebb2287d948b58c55c56ff5267164308e785e367"} Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.687562 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b" event={"ID":"b972200d-235d-4ae5-a71d-dd206e4f6025","Type":"ContainerStarted","Data":"ce7472e1afe221cc1944327c391ecdc90bb350dbee5e5d6abe5c9108a10db2f1"} Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.688496 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq" event={"ID":"df91f733-711d-43b0-98a0-b3fa6b119fbe","Type":"ContainerStarted","Data":"6a0466deaa419454e5de83d7e02e820d641a5c637a5d2f38732c9252330e4c3d"} Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.690221 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng" event={"ID":"f6e43ef6-1269-40f5-af4d-fe2f30dea319","Type":"ContainerStarted","Data":"f3b854e1b0db0de2a2989cbf10130964334c4ecf292a2741292d54e6034b0d6c"} Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.691613 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f" event={"ID":"5dd6bfb5-2edc-49b4-8877-85ab5722cdf0","Type":"ContainerStarted","Data":"a2781914a757c1b6efb9d567978e48fd0f730a91afed5b58bd3f32a274a0d319"} Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.692850 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm" event={"ID":"d3a93b19-a249-499a-89fc-9b2a1a69c1e0","Type":"ContainerStarted","Data":"7141fc5a3c8afd382569b4d8c30fda0610d65b2cabc642632be947ebacbb6462"} Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.767250 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c\" (UID: \"1c2d54cf-e53d-483d-9b00-6f03858806ab\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.767508 4857 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.767617 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert podName:1c2d54cf-e53d-483d-9b00-6f03858806ab nodeName:}" failed. No retries permitted until 2025-12-01 21:51:10.76754652 +0000 UTC m=+989.257608847 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" (UID: "1c2d54cf-e53d-483d-9b00-6f03858806ab") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.775244 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-cncd6"] Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.783280 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9"] Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.793158 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7"] Dec 01 21:51:08 crc kubenswrapper[4857]: W1201 21:51:08.800059 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod068a21df_001f_45e6_9946_68f67bad7d43.slice/crio-ac152e77cc587b8b7429b8bd8d3e2945fe4d2a5e4fff2faaa443522acde0f793 WatchSource:0}: Error finding container ac152e77cc587b8b7429b8bd8d3e2945fe4d2a5e4fff2faaa443522acde0f793: Status 404 returned error can't find the container with id ac152e77cc587b8b7429b8bd8d3e2945fe4d2a5e4fff2faaa443522acde0f793 Dec 01 21:51:08 crc kubenswrapper[4857]: W1201 21:51:08.800644 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcef9388_5d9b_4172_bb80_ec69b3e98dd0.slice/crio-edea9506fb87050d8e7b974ef00dfce677ba2cd39df1992c140aa04e37a11e6e WatchSource:0}: Error finding container edea9506fb87050d8e7b974ef00dfce677ba2cd39df1992c140aa04e37a11e6e: Status 404 returned error can't find the container with id edea9506fb87050d8e7b974ef00dfce677ba2cd39df1992c140aa04e37a11e6e Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.813259 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-mww86"] Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.824806 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887"] Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.835690 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d"] Dec 01 21:51:08 crc kubenswrapper[4857]: W1201 21:51:08.841454 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5ab847c_10d3_482d_8c70_e3ab439e7b5d.slice/crio-a21bb95d58aeb1f210d6d7da17786fc70f055f38c4e2e09e360ee37c31704d7a WatchSource:0}: Error finding container a21bb95d58aeb1f210d6d7da17786fc70f055f38c4e2e09e360ee37c31704d7a: Status 404 returned error can't find the container with id a21bb95d58aeb1f210d6d7da17786fc70f055f38c4e2e09e360ee37c31704d7a Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.843751 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9vwnb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-mww86_openstack-operators(c5ab847c-10d3-482d-8c70-e3ab439e7b5d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 21:51:08 crc kubenswrapper[4857]: W1201 21:51:08.845881 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff230e58_1711_4638_89d6_daf91fb3bfc1.slice/crio-853bc1366019ec62e51486d86dbb85da8f5b1359ca88fb89d521e53855005349 WatchSource:0}: Error finding container 853bc1366019ec62e51486d86dbb85da8f5b1359ca88fb89d521e53855005349: Status 404 returned error can't find the container with id 853bc1366019ec62e51486d86dbb85da8f5b1359ca88fb89d521e53855005349 Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.845933 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9vwnb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-mww86_openstack-operators(c5ab847c-10d3-482d-8c70-e3ab439e7b5d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.847030 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" podUID="c5ab847c-10d3-482d-8c70-e3ab439e7b5d" Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.849184 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.73:5001/openstack-k8s-operators/telemetry-operator:e82e6b4a488661603634ac58918e94b98a55620c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xjlql,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-7f765fc7d6-x797d_openstack-operators(ff230e58-1711-4638-89d6-daf91fb3bfc1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.851322 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xjlql,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-7f765fc7d6-x797d_openstack-operators(ff230e58-1711-4638-89d6-daf91fb3bfc1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.853208 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" podUID="ff230e58-1711-4638-89d6-daf91fb3bfc1" Dec 01 21:51:08 crc kubenswrapper[4857]: W1201 21:51:08.860597 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a6276de_e994_4f62_a625_906c9a5f4f5d.slice/crio-520c0e1bd66b53bdf4de32f8af1cb26356fe560f09b36f1ef309a5def029ea14 WatchSource:0}: Error finding container 520c0e1bd66b53bdf4de32f8af1cb26356fe560f09b36f1ef309a5def029ea14: Status 404 returned error can't find the container with id 520c0e1bd66b53bdf4de32f8af1cb26356fe560f09b36f1ef309a5def029ea14 Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.865499 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-96xzx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-vh887_openstack-operators(8a6276de-e994-4f62-a625-906c9a5f4f5d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.867511 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-96xzx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-vh887_openstack-operators(8a6276de-e994-4f62-a625-906c9a5f4f5d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.868769 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" podUID="8a6276de-e994-4f62-a625-906c9a5f4f5d" Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.919959 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wccvn"] Dec 01 21:51:08 crc kubenswrapper[4857]: W1201 21:51:08.923973 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c8d087e_e00c_4dad_87c7_06ab1017abaf.slice/crio-e1a920c14ef61801ec5f764c56b4f92b50013ae0958413abd44fe8b431ca17fe WatchSource:0}: Error finding container e1a920c14ef61801ec5f764c56b4f92b50013ae0958413abd44fe8b431ca17fe: Status 404 returned error can't find the container with id e1a920c14ef61801ec5f764c56b4f92b50013ae0958413abd44fe8b431ca17fe Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.942000 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v"] Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.945505 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-557w7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-p6v4v_openstack-operators(188a7d67-ffec-4608-9b5f-cc33972e6fed): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 21:51:08 crc kubenswrapper[4857]: W1201 21:51:08.946989 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcdd278ff_9eb6_45f7_977d_02944f6f312a.slice/crio-2754c856f4f648ec2f2a402b2ee9ceafe9ffbb9e1347f390ae20f61b5363416d WatchSource:0}: Error finding container 2754c856f4f648ec2f2a402b2ee9ceafe9ffbb9e1347f390ae20f61b5363416d: Status 404 returned error can't find the container with id 2754c856f4f648ec2f2a402b2ee9ceafe9ffbb9e1347f390ae20f61b5363416d Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.947480 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-557w7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-p6v4v_openstack-operators(188a7d67-ffec-4608-9b5f-cc33972e6fed): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.948563 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" podUID="188a7d67-ffec-4608-9b5f-cc33972e6fed" Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.949118 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-47g5g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-wjnzv_openstack-operators(cdd278ff-9eb6-45f7-977d-02944f6f312a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.955457 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv"] Dec 01 21:51:08 crc kubenswrapper[4857]: I1201 21:51:08.961377 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh"] Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.978596 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-47g5g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-wjnzv_openstack-operators(cdd278ff-9eb6-45f7-977d-02944f6f312a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.979778 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" podUID="cdd278ff-9eb6-45f7-977d-02944f6f312a" Dec 01 21:51:08 crc kubenswrapper[4857]: E1201 21:51:08.999032 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ndrc2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-gdqjh_openstack-operators(1d490d1b-558b-49fb-9b0f-65d0f82b0148): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 21:51:09 crc kubenswrapper[4857]: E1201 21:51:09.000303 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh" podUID="1d490d1b-558b-49fb-9b0f-65d0f82b0148" Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.176377 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.176494 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:09 crc kubenswrapper[4857]: E1201 21:51:09.176616 4857 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 21:51:09 crc kubenswrapper[4857]: E1201 21:51:09.176654 4857 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 21:51:09 crc kubenswrapper[4857]: E1201 21:51:09.176670 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs podName:0a4c6432-bb6c-41b6-9981-ef6ac1354ae3 nodeName:}" failed. No retries permitted until 2025-12-01 21:51:11.176655778 +0000 UTC m=+989.666718095 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs") pod "openstack-operator-controller-manager-777b97d487-s7g6x" (UID: "0a4c6432-bb6c-41b6-9981-ef6ac1354ae3") : secret "metrics-server-cert" not found Dec 01 21:51:09 crc kubenswrapper[4857]: E1201 21:51:09.176739 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs podName:0a4c6432-bb6c-41b6-9981-ef6ac1354ae3 nodeName:}" failed. No retries permitted until 2025-12-01 21:51:11.176713959 +0000 UTC m=+989.666776366 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs") pod "openstack-operator-controller-manager-777b97d487-s7g6x" (UID: "0a4c6432-bb6c-41b6-9981-ef6ac1354ae3") : secret "webhook-server-cert" not found Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.743117 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" event={"ID":"ff230e58-1711-4638-89d6-daf91fb3bfc1","Type":"ContainerStarted","Data":"853bc1366019ec62e51486d86dbb85da8f5b1359ca88fb89d521e53855005349"} Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.746215 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cncd6" event={"ID":"246a4bf0-515b-46f8-bd7f-89e7e961d774","Type":"ContainerStarted","Data":"6e8cacfe51759f10ca3fb86e8c7297962944d251c400dbfa72763b40a38567da"} Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.747608 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7" event={"ID":"dcef9388-5d9b-4172-bb80-ec69b3e98dd0","Type":"ContainerStarted","Data":"edea9506fb87050d8e7b974ef00dfce677ba2cd39df1992c140aa04e37a11e6e"} Dec 01 21:51:09 crc kubenswrapper[4857]: E1201 21:51:09.748638 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.73:5001/openstack-k8s-operators/telemetry-operator:e82e6b4a488661603634ac58918e94b98a55620c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" podUID="ff230e58-1711-4638-89d6-daf91fb3bfc1" Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.749206 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh" event={"ID":"1d490d1b-558b-49fb-9b0f-65d0f82b0148","Type":"ContainerStarted","Data":"14c98c904591ad1f9f1704010798b5393ece504e71c6d3ddd03f318f11ad4c27"} Dec 01 21:51:09 crc kubenswrapper[4857]: E1201 21:51:09.751465 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh" podUID="1d490d1b-558b-49fb-9b0f-65d0f82b0148" Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.752392 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wccvn" event={"ID":"4c8d087e-e00c-4dad-87c7-06ab1017abaf","Type":"ContainerStarted","Data":"e1a920c14ef61801ec5f764c56b4f92b50013ae0958413abd44fe8b431ca17fe"} Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.754251 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb" event={"ID":"92c43db5-8128-4269-9ad4-a263de88f450","Type":"ContainerStarted","Data":"f685abfdcde35be51a22ba32cfe00b66d590761c7647cfec40fc354b89114552"} Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.755866 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" event={"ID":"188a7d67-ffec-4608-9b5f-cc33972e6fed","Type":"ContainerStarted","Data":"ce9d2e4ece50706cd76e9639a3c8902a55420310af2e612fbb8c415cd327cc8a"} Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.757624 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" event={"ID":"c5ab847c-10d3-482d-8c70-e3ab439e7b5d","Type":"ContainerStarted","Data":"a21bb95d58aeb1f210d6d7da17786fc70f055f38c4e2e09e360ee37c31704d7a"} Dec 01 21:51:09 crc kubenswrapper[4857]: E1201 21:51:09.759420 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" podUID="188a7d67-ffec-4608-9b5f-cc33972e6fed" Dec 01 21:51:09 crc kubenswrapper[4857]: E1201 21:51:09.760315 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" podUID="c5ab847c-10d3-482d-8c70-e3ab439e7b5d" Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.760726 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" event={"ID":"cdd278ff-9eb6-45f7-977d-02944f6f312a","Type":"ContainerStarted","Data":"2754c856f4f648ec2f2a402b2ee9ceafe9ffbb9e1347f390ae20f61b5363416d"} Dec 01 21:51:09 crc kubenswrapper[4857]: E1201 21:51:09.762317 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" podUID="cdd278ff-9eb6-45f7-977d-02944f6f312a" Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.762473 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" event={"ID":"8a6276de-e994-4f62-a625-906c9a5f4f5d","Type":"ContainerStarted","Data":"520c0e1bd66b53bdf4de32f8af1cb26356fe560f09b36f1ef309a5def029ea14"} Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.764169 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt" event={"ID":"de53c246-6e20-4b67-be68-1c5356fd46d4","Type":"ContainerStarted","Data":"0887196987f7b5360283dfb1f9b821f762e6bcee893a646129669f784c86609f"} Dec 01 21:51:09 crc kubenswrapper[4857]: E1201 21:51:09.765558 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" podUID="8a6276de-e994-4f62-a625-906c9a5f4f5d" Dec 01 21:51:09 crc kubenswrapper[4857]: I1201 21:51:09.765758 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9" event={"ID":"068a21df-001f-45e6-9946-68f67bad7d43","Type":"ContainerStarted","Data":"ac152e77cc587b8b7429b8bd8d3e2945fe4d2a5e4fff2faaa443522acde0f793"} Dec 01 21:51:10 crc kubenswrapper[4857]: I1201 21:51:10.251179 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert\") pod \"infra-operator-controller-manager-57548d458d-54xtc\" (UID: \"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:10 crc kubenswrapper[4857]: E1201 21:51:10.251335 4857 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 21:51:10 crc kubenswrapper[4857]: E1201 21:51:10.251389 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert podName:23e7348c-65d9-4c4b-9c1e-d86cc3b55bff nodeName:}" failed. No retries permitted until 2025-12-01 21:51:14.251371469 +0000 UTC m=+992.741433786 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert") pod "infra-operator-controller-manager-57548d458d-54xtc" (UID: "23e7348c-65d9-4c4b-9c1e-d86cc3b55bff") : secret "infra-operator-webhook-server-cert" not found Dec 01 21:51:10 crc kubenswrapper[4857]: E1201 21:51:10.789079 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh" podUID="1d490d1b-558b-49fb-9b0f-65d0f82b0148" Dec 01 21:51:10 crc kubenswrapper[4857]: E1201 21:51:10.790666 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.73:5001/openstack-k8s-operators/telemetry-operator:e82e6b4a488661603634ac58918e94b98a55620c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" podUID="ff230e58-1711-4638-89d6-daf91fb3bfc1" Dec 01 21:51:10 crc kubenswrapper[4857]: E1201 21:51:10.791031 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" podUID="cdd278ff-9eb6-45f7-977d-02944f6f312a" Dec 01 21:51:10 crc kubenswrapper[4857]: E1201 21:51:10.791855 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" podUID="188a7d67-ffec-4608-9b5f-cc33972e6fed" Dec 01 21:51:10 crc kubenswrapper[4857]: E1201 21:51:10.791924 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" podUID="8a6276de-e994-4f62-a625-906c9a5f4f5d" Dec 01 21:51:10 crc kubenswrapper[4857]: E1201 21:51:10.792434 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" podUID="c5ab847c-10d3-482d-8c70-e3ab439e7b5d" Dec 01 21:51:10 crc kubenswrapper[4857]: I1201 21:51:10.811550 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c\" (UID: \"1c2d54cf-e53d-483d-9b00-6f03858806ab\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:10 crc kubenswrapper[4857]: E1201 21:51:10.813470 4857 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 21:51:10 crc kubenswrapper[4857]: E1201 21:51:10.813545 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert podName:1c2d54cf-e53d-483d-9b00-6f03858806ab nodeName:}" failed. No retries permitted until 2025-12-01 21:51:14.81352669 +0000 UTC m=+993.303589007 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" (UID: "1c2d54cf-e53d-483d-9b00-6f03858806ab") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 21:51:11 crc kubenswrapper[4857]: I1201 21:51:11.201029 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:11 crc kubenswrapper[4857]: I1201 21:51:11.201211 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:11 crc kubenswrapper[4857]: E1201 21:51:11.201419 4857 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 21:51:11 crc kubenswrapper[4857]: E1201 21:51:11.201534 4857 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 21:51:11 crc kubenswrapper[4857]: E1201 21:51:11.201577 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs podName:0a4c6432-bb6c-41b6-9981-ef6ac1354ae3 nodeName:}" failed. No retries permitted until 2025-12-01 21:51:15.201481018 +0000 UTC m=+993.691543335 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs") pod "openstack-operator-controller-manager-777b97d487-s7g6x" (UID: "0a4c6432-bb6c-41b6-9981-ef6ac1354ae3") : secret "metrics-server-cert" not found Dec 01 21:51:11 crc kubenswrapper[4857]: E1201 21:51:11.201848 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs podName:0a4c6432-bb6c-41b6-9981-ef6ac1354ae3 nodeName:}" failed. No retries permitted until 2025-12-01 21:51:15.201820156 +0000 UTC m=+993.691882473 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs") pod "openstack-operator-controller-manager-777b97d487-s7g6x" (UID: "0a4c6432-bb6c-41b6-9981-ef6ac1354ae3") : secret "webhook-server-cert" not found Dec 01 21:51:14 crc kubenswrapper[4857]: I1201 21:51:14.297564 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert\") pod \"infra-operator-controller-manager-57548d458d-54xtc\" (UID: \"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:14 crc kubenswrapper[4857]: E1201 21:51:14.297776 4857 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 21:51:14 crc kubenswrapper[4857]: E1201 21:51:14.297971 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert podName:23e7348c-65d9-4c4b-9c1e-d86cc3b55bff nodeName:}" failed. No retries permitted until 2025-12-01 21:51:22.297953173 +0000 UTC m=+1000.788015490 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert") pod "infra-operator-controller-manager-57548d458d-54xtc" (UID: "23e7348c-65d9-4c4b-9c1e-d86cc3b55bff") : secret "infra-operator-webhook-server-cert" not found Dec 01 21:51:14 crc kubenswrapper[4857]: I1201 21:51:14.948057 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c\" (UID: \"1c2d54cf-e53d-483d-9b00-6f03858806ab\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:14 crc kubenswrapper[4857]: E1201 21:51:14.948261 4857 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 21:51:14 crc kubenswrapper[4857]: E1201 21:51:14.948311 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert podName:1c2d54cf-e53d-483d-9b00-6f03858806ab nodeName:}" failed. No retries permitted until 2025-12-01 21:51:22.948295751 +0000 UTC m=+1001.438358078 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" (UID: "1c2d54cf-e53d-483d-9b00-6f03858806ab") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 21:51:15 crc kubenswrapper[4857]: I1201 21:51:15.269163 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:15 crc kubenswrapper[4857]: I1201 21:51:15.269288 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:15 crc kubenswrapper[4857]: E1201 21:51:15.269352 4857 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 21:51:15 crc kubenswrapper[4857]: E1201 21:51:15.269436 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs podName:0a4c6432-bb6c-41b6-9981-ef6ac1354ae3 nodeName:}" failed. No retries permitted until 2025-12-01 21:51:23.269417995 +0000 UTC m=+1001.759480312 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs") pod "openstack-operator-controller-manager-777b97d487-s7g6x" (UID: "0a4c6432-bb6c-41b6-9981-ef6ac1354ae3") : secret "metrics-server-cert" not found Dec 01 21:51:15 crc kubenswrapper[4857]: E1201 21:51:15.269463 4857 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 21:51:15 crc kubenswrapper[4857]: E1201 21:51:15.269525 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs podName:0a4c6432-bb6c-41b6-9981-ef6ac1354ae3 nodeName:}" failed. No retries permitted until 2025-12-01 21:51:23.269510438 +0000 UTC m=+1001.759572755 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs") pod "openstack-operator-controller-manager-777b97d487-s7g6x" (UID: "0a4c6432-bb6c-41b6-9981-ef6ac1354ae3") : secret "webhook-server-cert" not found Dec 01 21:51:22 crc kubenswrapper[4857]: I1201 21:51:22.303751 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert\") pod \"infra-operator-controller-manager-57548d458d-54xtc\" (UID: \"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:22 crc kubenswrapper[4857]: I1201 21:51:22.320127 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23e7348c-65d9-4c4b-9c1e-d86cc3b55bff-cert\") pod \"infra-operator-controller-manager-57548d458d-54xtc\" (UID: \"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:22 crc kubenswrapper[4857]: I1201 21:51:22.604362 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:23 crc kubenswrapper[4857]: I1201 21:51:23.013069 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c\" (UID: \"1c2d54cf-e53d-483d-9b00-6f03858806ab\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:23 crc kubenswrapper[4857]: I1201 21:51:23.018414 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c2d54cf-e53d-483d-9b00-6f03858806ab-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c\" (UID: \"1c2d54cf-e53d-483d-9b00-6f03858806ab\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:23 crc kubenswrapper[4857]: I1201 21:51:23.285715 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:23 crc kubenswrapper[4857]: I1201 21:51:23.316827 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:23 crc kubenswrapper[4857]: I1201 21:51:23.316898 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:23 crc kubenswrapper[4857]: I1201 21:51:23.328797 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-metrics-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:23 crc kubenswrapper[4857]: I1201 21:51:23.329182 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0a4c6432-bb6c-41b6-9981-ef6ac1354ae3-webhook-certs\") pod \"openstack-operator-controller-manager-777b97d487-s7g6x\" (UID: \"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3\") " pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:23 crc kubenswrapper[4857]: I1201 21:51:23.629233 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:26 crc kubenswrapper[4857]: E1201 21:51:26.241430 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:ecf7be921850bdc04697ed1b332bab39ad2a64e4e45c2a445c04f9bae6ac61b5" Dec 01 21:51:26 crc kubenswrapper[4857]: E1201 21:51:26.242308 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:ecf7be921850bdc04697ed1b332bab39ad2a64e4e45c2a445c04f9bae6ac61b5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x6f7l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6546668bfd-b7ksb_openstack-operators(92c43db5-8128-4269-9ad4-a263de88f450): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:51:26 crc kubenswrapper[4857]: E1201 21:51:26.967803 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 01 21:51:26 crc kubenswrapper[4857]: E1201 21:51:26.967981 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wvgg5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-rqfzk_openstack-operators(80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:51:27 crc kubenswrapper[4857]: I1201 21:51:27.805790 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:51:27 crc kubenswrapper[4857]: I1201 21:51:27.806220 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:51:28 crc kubenswrapper[4857]: E1201 21:51:28.357114 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 01 21:51:28 crc kubenswrapper[4857]: E1201 21:51:28.357397 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mp7nf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-zkqw9_openstack-operators(068a21df-001f-45e6-9946-68f67bad7d43): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:51:29 crc kubenswrapper[4857]: E1201 21:51:29.031142 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 01 21:51:29 crc kubenswrapper[4857]: E1201 21:51:29.031374 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-62vjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-f9d2t_openstack-operators(b13a13cc-0689-480a-bc95-e19aba327f9c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:51:29 crc kubenswrapper[4857]: E1201 21:51:29.818634 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 01 21:51:29 crc kubenswrapper[4857]: E1201 21:51:29.818856 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gmfpc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-7vmng_openstack-operators(f6e43ef6-1269-40f5-af4d-fe2f30dea319): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:51:30 crc kubenswrapper[4857]: E1201 21:51:30.398797 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 01 21:51:30 crc kubenswrapper[4857]: E1201 21:51:30.399529 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m7mgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-j277f_openstack-operators(5dd6bfb5-2edc-49b4-8877-85ab5722cdf0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:51:30 crc kubenswrapper[4857]: E1201 21:51:30.941723 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 01 21:51:30 crc kubenswrapper[4857]: E1201 21:51:30.942030 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6hp7q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-wccvn_openstack-operators(4c8d087e-e00c-4dad-87c7-06ab1017abaf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:51:31 crc kubenswrapper[4857]: E1201 21:51:31.434886 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 01 21:51:31 crc kubenswrapper[4857]: E1201 21:51:31.435151 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-28nhf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-vbvnt_openstack-operators(de53c246-6e20-4b67-be68-1c5356fd46d4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:51:32 crc kubenswrapper[4857]: E1201 21:51:32.095538 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3" Dec 01 21:51:32 crc kubenswrapper[4857]: E1201 21:51:32.095759 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lqmqx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-546d4bdf48-vtcx7_openstack-operators(dcef9388-5d9b-4172-bb80-ec69b3e98dd0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:51:39 crc kubenswrapper[4857]: I1201 21:51:39.228637 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c"] Dec 01 21:51:39 crc kubenswrapper[4857]: I1201 21:51:39.308091 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x"] Dec 01 21:51:39 crc kubenswrapper[4857]: I1201 21:51:39.317928 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-54xtc"] Dec 01 21:51:39 crc kubenswrapper[4857]: W1201 21:51:39.525304 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c2d54cf_e53d_483d_9b00_6f03858806ab.slice/crio-f3f15f2b0937ef6288b2038b40fe4ca234c9dc2478998d2ac0dead648c5aba01 WatchSource:0}: Error finding container f3f15f2b0937ef6288b2038b40fe4ca234c9dc2478998d2ac0dead648c5aba01: Status 404 returned error can't find the container with id f3f15f2b0937ef6288b2038b40fe4ca234c9dc2478998d2ac0dead648c5aba01 Dec 01 21:51:39 crc kubenswrapper[4857]: I1201 21:51:39.528343 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 21:51:40 crc kubenswrapper[4857]: I1201 21:51:40.230273 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cncd6" event={"ID":"246a4bf0-515b-46f8-bd7f-89e7e961d774","Type":"ContainerStarted","Data":"7b3168f3e97cbcadb4a2ee9f7c72f6547b7fb2acc7a014fce81faaca56e17c96"} Dec 01 21:51:40 crc kubenswrapper[4857]: I1201 21:51:40.231932 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" event={"ID":"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3","Type":"ContainerStarted","Data":"6daa4d9bc3692edb45e6da249b806be18818a0a305a76bbe47ffbdb47c4dbc9d"} Dec 01 21:51:40 crc kubenswrapper[4857]: I1201 21:51:40.235388 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" event={"ID":"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff","Type":"ContainerStarted","Data":"f569156443f3412c8952fd03ecef3efb50395216eac365cbfdce7f18fe7dc716"} Dec 01 21:51:40 crc kubenswrapper[4857]: I1201 21:51:40.236951 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b" event={"ID":"b972200d-235d-4ae5-a71d-dd206e4f6025","Type":"ContainerStarted","Data":"b9d67ed201e35062f183bc53a9820ce6bf10da531d838932cc9729cf412dae65"} Dec 01 21:51:40 crc kubenswrapper[4857]: I1201 21:51:40.238396 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq" event={"ID":"df91f733-711d-43b0-98a0-b3fa6b119fbe","Type":"ContainerStarted","Data":"288ac9b064e890bcb12bbc1e536881ac104653809c16fc817adb91eaefb9a7bd"} Dec 01 21:51:40 crc kubenswrapper[4857]: I1201 21:51:40.239480 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" event={"ID":"1c2d54cf-e53d-483d-9b00-6f03858806ab","Type":"ContainerStarted","Data":"f3f15f2b0937ef6288b2038b40fe4ca234c9dc2478998d2ac0dead648c5aba01"} Dec 01 21:51:41 crc kubenswrapper[4857]: I1201 21:51:41.275917 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" event={"ID":"c5ab847c-10d3-482d-8c70-e3ab439e7b5d","Type":"ContainerStarted","Data":"3cd0c434d1c845bb801c50bb58955361ea6d15db5d2f432ba8f941054937730c"} Dec 01 21:51:41 crc kubenswrapper[4857]: I1201 21:51:41.282118 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" event={"ID":"cdd278ff-9eb6-45f7-977d-02944f6f312a","Type":"ContainerStarted","Data":"ec7af30a69dcf76a00ba2592eadaf9af533186e4e554f81c1063656a2fe8d998"} Dec 01 21:51:41 crc kubenswrapper[4857]: I1201 21:51:41.284139 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" event={"ID":"8a6276de-e994-4f62-a625-906c9a5f4f5d","Type":"ContainerStarted","Data":"b4b03ac278caa4a86b966bd9b6f8ce9aa576e21f7201a0bafe4fbcd290d6e0db"} Dec 01 21:51:41 crc kubenswrapper[4857]: I1201 21:51:41.287154 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm" event={"ID":"d3a93b19-a249-499a-89fc-9b2a1a69c1e0","Type":"ContainerStarted","Data":"40a0fcd969229aee965740d9eeca8b34f079719fe237e27c0ca461f3febdddcf"} Dec 01 21:51:41 crc kubenswrapper[4857]: I1201 21:51:41.288446 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" event={"ID":"188a7d67-ffec-4608-9b5f-cc33972e6fed","Type":"ContainerStarted","Data":"241ed68def8d2982ccf83ec0b245f6f7427fcfcce7bb5991a0e7981ada6886ef"} Dec 01 21:51:41 crc kubenswrapper[4857]: I1201 21:51:41.289791 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5" event={"ID":"393fd1a5-e3b9-44d6-a9ef-1447a0f9f465","Type":"ContainerStarted","Data":"ee0259b9fe96ee5671b29361d740eff906394293d3089c48a595e9d92d54b396"} Dec 01 21:51:42 crc kubenswrapper[4857]: I1201 21:51:42.310882 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh" event={"ID":"1d490d1b-558b-49fb-9b0f-65d0f82b0148","Type":"ContainerStarted","Data":"8befeaad682c53bf4fe552b69476d45d17d4dde98fd8591ba23c32319d7da88f"} Dec 01 21:51:42 crc kubenswrapper[4857]: I1201 21:51:42.316696 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" event={"ID":"0a4c6432-bb6c-41b6-9981-ef6ac1354ae3","Type":"ContainerStarted","Data":"81d15fc35b1254ac3b381f168d0e6ec62b3fbb42cd1ab95bdf536ed95038872c"} Dec 01 21:51:42 crc kubenswrapper[4857]: I1201 21:51:42.316908 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:42 crc kubenswrapper[4857]: I1201 21:51:42.330885 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gdqjh" podStartSLOduration=5.397359236 podStartE2EDuration="35.330870532s" podCreationTimestamp="2025-12-01 21:51:07 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.998898377 +0000 UTC m=+987.488960694" lastFinishedPulling="2025-12-01 21:51:38.932409673 +0000 UTC m=+1017.422471990" observedRunningTime="2025-12-01 21:51:42.325057139 +0000 UTC m=+1020.815119456" watchObservedRunningTime="2025-12-01 21:51:42.330870532 +0000 UTC m=+1020.820932839" Dec 01 21:51:48 crc kubenswrapper[4857]: I1201 21:51:48.358827 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" event={"ID":"ff230e58-1711-4638-89d6-daf91fb3bfc1","Type":"ContainerStarted","Data":"af49e5c465c21069cf767815e5acd3be08c77f334aca4b9f97608d2ca7428859"} Dec 01 21:51:49 crc kubenswrapper[4857]: I1201 21:51:49.371119 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" event={"ID":"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff","Type":"ContainerStarted","Data":"e85e25b7927655b1fa7a322c62cd64118fc38bfdb36e9137c42d69c7c0d7db79"} Dec 01 21:51:49 crc kubenswrapper[4857]: E1201 21:51:49.500901 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng" podUID="f6e43ef6-1269-40f5-af4d-fe2f30dea319" Dec 01 21:51:49 crc kubenswrapper[4857]: E1201 21:51:49.558290 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wccvn" podUID="4c8d087e-e00c-4dad-87c7-06ab1017abaf" Dec 01 21:51:49 crc kubenswrapper[4857]: E1201 21:51:49.602527 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb" podUID="92c43db5-8128-4269-9ad4-a263de88f450" Dec 01 21:51:49 crc kubenswrapper[4857]: E1201 21:51:49.664952 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt" podUID="de53c246-6e20-4b67-be68-1c5356fd46d4" Dec 01 21:51:49 crc kubenswrapper[4857]: E1201 21:51:49.712655 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7" podUID="dcef9388-5d9b-4172-bb80-ec69b3e98dd0" Dec 01 21:51:49 crc kubenswrapper[4857]: E1201 21:51:49.712992 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f" podUID="5dd6bfb5-2edc-49b4-8877-85ab5722cdf0" Dec 01 21:51:49 crc kubenswrapper[4857]: E1201 21:51:49.739575 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk" podUID="80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4" Dec 01 21:51:49 crc kubenswrapper[4857]: E1201 21:51:49.938728 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t" podUID="b13a13cc-0689-480a-bc95-e19aba327f9c" Dec 01 21:51:50 crc kubenswrapper[4857]: E1201 21:51:50.004726 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9" podUID="068a21df-001f-45e6-9946-68f67bad7d43" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.378263 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" event={"ID":"23e7348c-65d9-4c4b-9c1e-d86cc3b55bff","Type":"ContainerStarted","Data":"84a6dfda6d013547b236a413c90602d6b75e71f1b5a6b7646afe45099753e896"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.378582 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.379770 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt" event={"ID":"de53c246-6e20-4b67-be68-1c5356fd46d4","Type":"ContainerStarted","Data":"90bae3314be52613fe51f334954ced6b56a43cfed340af80ced98306fb181f3f"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.382218 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9" event={"ID":"068a21df-001f-45e6-9946-68f67bad7d43","Type":"ContainerStarted","Data":"f552d79339aef1597f3deb0f84b7dbc6a99e0c7f3e67a1d27310bcd96b97f7c3"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.384405 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b" event={"ID":"b972200d-235d-4ae5-a71d-dd206e4f6025","Type":"ContainerStarted","Data":"866e694706efb539154b84cce299077994022f384b096fc008b7775e32242de6"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.385014 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.387169 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.389121 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" event={"ID":"188a7d67-ffec-4608-9b5f-cc33972e6fed","Type":"ContainerStarted","Data":"be9826364e2fc07526386401245c36ff4d18fba8df56d04ed5ae888ac8b666dc"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.389338 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.391244 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng" event={"ID":"f6e43ef6-1269-40f5-af4d-fe2f30dea319","Type":"ContainerStarted","Data":"510122bff667f3d77bbfca15726bfafba111cc3ac44198dc1d765a45fd562d9b"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.392489 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f" event={"ID":"5dd6bfb5-2edc-49b4-8877-85ab5722cdf0","Type":"ContainerStarted","Data":"a2e70500929776a6542e26d7a35ee5a70ee513e84fe49928f295e3c0f1629bee"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.392861 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.394269 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wccvn" event={"ID":"4c8d087e-e00c-4dad-87c7-06ab1017abaf","Type":"ContainerStarted","Data":"fe0f932a41aa6792ca4d72ab5661ab85f8683b4041cb1b76c463ed4adca6bad6"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.396081 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm" event={"ID":"d3a93b19-a249-499a-89fc-9b2a1a69c1e0","Type":"ContainerStarted","Data":"4051e9bdfc6680adb15afd719ab1453ab34e957f3fa87e930a1956d78c6504ce"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.396898 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.405278 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7" event={"ID":"dcef9388-5d9b-4172-bb80-ec69b3e98dd0","Type":"ContainerStarted","Data":"574a969b347e0a1f98289f1c5b8adc0e819365892a6d77c8ba92c99df034c791"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.407500 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" podStartSLOduration=43.407490073 podStartE2EDuration="43.407490073s" podCreationTimestamp="2025-12-01 21:51:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:51:42.366895058 +0000 UTC m=+1020.856957375" watchObservedRunningTime="2025-12-01 21:51:50.407490073 +0000 UTC m=+1028.897552390" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.411922 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" podStartSLOduration=35.825304234 podStartE2EDuration="44.411908882s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:39.720027086 +0000 UTC m=+1018.210089403" lastFinishedPulling="2025-12-01 21:51:48.306631734 +0000 UTC m=+1026.796694051" observedRunningTime="2025-12-01 21:51:50.404889919 +0000 UTC m=+1028.894952236" watchObservedRunningTime="2025-12-01 21:51:50.411908882 +0000 UTC m=+1028.901971199" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.415267 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" event={"ID":"1c2d54cf-e53d-483d-9b00-6f03858806ab","Type":"ContainerStarted","Data":"e6913236b87619b80f755024a8e39f85cfc6e54757a764f0e91dc47d66c3745b"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.415328 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" event={"ID":"1c2d54cf-e53d-483d-9b00-6f03858806ab","Type":"ContainerStarted","Data":"18a27c8052603125f8b1301a3d15900c43a5bbc32f6a03cf83e9af3ece933864"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.416413 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.416951 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.431170 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cncd6" event={"ID":"246a4bf0-515b-46f8-bd7f-89e7e961d774","Type":"ContainerStarted","Data":"b9429ec550a4d9aecb6b670e8f83a933461429de37f41d745e85fb82d7c0c8e5"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.433102 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cncd6" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.437367 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cncd6" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.455279 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk" event={"ID":"80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4","Type":"ContainerStarted","Data":"dbff64a082d1d83b513a6430b37a55b28e478df1ebf02d7391611c3eb541a177"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.470936 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5" event={"ID":"393fd1a5-e3b9-44d6-a9ef-1447a0f9f465","Type":"ContainerStarted","Data":"c7e347c6817bb715197050b08cf382791cec6d20ffee5e5295509771cce11b05"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.472130 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.481317 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.489591 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-5gqcm" podStartSLOduration=3.676205929 podStartE2EDuration="44.489569411s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:07.962213521 +0000 UTC m=+986.452275838" lastFinishedPulling="2025-12-01 21:51:48.775576963 +0000 UTC m=+1027.265639320" observedRunningTime="2025-12-01 21:51:50.454052498 +0000 UTC m=+1028.944114815" watchObservedRunningTime="2025-12-01 21:51:50.489569411 +0000 UTC m=+1028.979631728" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.490566 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" event={"ID":"c5ab847c-10d3-482d-8c70-e3ab439e7b5d","Type":"ContainerStarted","Data":"a48fedc6d7eb7b645f9acdf8a66793fe3e24f757b0d94b42575c4a2a899ceafb"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.491789 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.496155 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.496729 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p6v4v" podStartSLOduration=4.66427358 podStartE2EDuration="44.496703896s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.94533213 +0000 UTC m=+987.435394447" lastFinishedPulling="2025-12-01 21:51:48.777762446 +0000 UTC m=+1027.267824763" observedRunningTime="2025-12-01 21:51:50.484695652 +0000 UTC m=+1028.974757979" watchObservedRunningTime="2025-12-01 21:51:50.496703896 +0000 UTC m=+1028.986766213" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.507273 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" event={"ID":"ff230e58-1711-4638-89d6-daf91fb3bfc1","Type":"ContainerStarted","Data":"0bae107d09680e03ea4614955fa9209a5271afc82dba036ee1b8e2f85c481bad"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.507996 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.515476 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" event={"ID":"cdd278ff-9eb6-45f7-977d-02944f6f312a","Type":"ContainerStarted","Data":"36f8d41f30543fec8fcd9aac4072a783e0f01dc309972cca8e2f6b5a5fcb5084"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.517820 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.518356 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.519178 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t" event={"ID":"b13a13cc-0689-480a-bc95-e19aba327f9c","Type":"ContainerStarted","Data":"ece441c114eb72ece70aa0ace981a8a28862a76e545d5e106816096bf2012757"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.550157 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq" event={"ID":"df91f733-711d-43b0-98a0-b3fa6b119fbe","Type":"ContainerStarted","Data":"95df379e2221537dfd8eac4c5ee2a6ab8c81fda008dce4dad66a273c967c0465"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.550268 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.552491 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.597246 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb" event={"ID":"92c43db5-8128-4269-9ad4-a263de88f450","Type":"ContainerStarted","Data":"6cf46f528ba1bba9afa7467497681e44385ab0e8b49386c89f23c9189e902064"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.616076 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4g68b" podStartSLOduration=4.175777961 podStartE2EDuration="44.616058431s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.327731057 +0000 UTC m=+986.817793374" lastFinishedPulling="2025-12-01 21:51:48.768011517 +0000 UTC m=+1027.258073844" observedRunningTime="2025-12-01 21:51:50.604325562 +0000 UTC m=+1029.094387879" watchObservedRunningTime="2025-12-01 21:51:50.616058431 +0000 UTC m=+1029.106120748" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.619947 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" event={"ID":"8a6276de-e994-4f62-a625-906c9a5f4f5d","Type":"ContainerStarted","Data":"531431dc73b5781e557038edde986ad76857199b53e4ed2d16c69a001b187054"} Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.620981 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.641066 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.753724 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mww86" podStartSLOduration=4.820634375 podStartE2EDuration="44.753700305s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.84363545 +0000 UTC m=+987.333697767" lastFinishedPulling="2025-12-01 21:51:48.77670134 +0000 UTC m=+1027.266763697" observedRunningTime="2025-12-01 21:51:50.744194181 +0000 UTC m=+1029.234256498" watchObservedRunningTime="2025-12-01 21:51:50.753700305 +0000 UTC m=+1029.243762622" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.803213 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" podStartSLOduration=36.024728258 podStartE2EDuration="44.803199622s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:39.528107968 +0000 UTC m=+1018.018170285" lastFinishedPulling="2025-12-01 21:51:48.306579332 +0000 UTC m=+1026.796641649" observedRunningTime="2025-12-01 21:51:50.799415978 +0000 UTC m=+1029.289478295" watchObservedRunningTime="2025-12-01 21:51:50.803199622 +0000 UTC m=+1029.293261939" Dec 01 21:51:50 crc kubenswrapper[4857]: I1201 21:51:50.919717 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-dz5k5" podStartSLOduration=4.556239724 podStartE2EDuration="44.919698065s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.416014187 +0000 UTC m=+986.906076504" lastFinishedPulling="2025-12-01 21:51:48.779472528 +0000 UTC m=+1027.269534845" observedRunningTime="2025-12-01 21:51:50.91789878 +0000 UTC m=+1029.407961097" watchObservedRunningTime="2025-12-01 21:51:50.919698065 +0000 UTC m=+1029.409760392" Dec 01 21:51:51 crc kubenswrapper[4857]: I1201 21:51:51.012509 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cncd6" podStartSLOduration=4.956627577 podStartE2EDuration="45.012496436s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.779648777 +0000 UTC m=+987.269711094" lastFinishedPulling="2025-12-01 21:51:48.835517616 +0000 UTC m=+1027.325579953" observedRunningTime="2025-12-01 21:51:51.009567804 +0000 UTC m=+1029.499630121" watchObservedRunningTime="2025-12-01 21:51:51.012496436 +0000 UTC m=+1029.502558743" Dec 01 21:51:51 crc kubenswrapper[4857]: I1201 21:51:51.036778 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" podStartSLOduration=13.933866549 podStartE2EDuration="45.036761143s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.849061234 +0000 UTC m=+987.339123551" lastFinishedPulling="2025-12-01 21:51:39.951955838 +0000 UTC m=+1018.442018145" observedRunningTime="2025-12-01 21:51:51.031733839 +0000 UTC m=+1029.521796156" watchObservedRunningTime="2025-12-01 21:51:51.036761143 +0000 UTC m=+1029.526823460" Dec 01 21:51:51 crc kubenswrapper[4857]: I1201 21:51:51.056450 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vh887" podStartSLOduration=5.144349482 podStartE2EDuration="45.056432456s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.865306373 +0000 UTC m=+987.355368700" lastFinishedPulling="2025-12-01 21:51:48.777389357 +0000 UTC m=+1027.267451674" observedRunningTime="2025-12-01 21:51:51.053275239 +0000 UTC m=+1029.543337556" watchObservedRunningTime="2025-12-01 21:51:51.056432456 +0000 UTC m=+1029.546494773" Dec 01 21:51:51 crc kubenswrapper[4857]: I1201 21:51:51.077684 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-wjnzv" podStartSLOduration=5.210877658 podStartE2EDuration="45.077669208s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.94896129 +0000 UTC m=+987.439023607" lastFinishedPulling="2025-12-01 21:51:48.81575283 +0000 UTC m=+1027.305815157" observedRunningTime="2025-12-01 21:51:51.074366427 +0000 UTC m=+1029.564428744" watchObservedRunningTime="2025-12-01 21:51:51.077669208 +0000 UTC m=+1029.567731525" Dec 01 21:51:51 crc kubenswrapper[4857]: I1201 21:51:51.138718 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-9qwbq" podStartSLOduration=4.748085301 podStartE2EDuration="45.138697019s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.332604237 +0000 UTC m=+986.822666554" lastFinishedPulling="2025-12-01 21:51:48.723215945 +0000 UTC m=+1027.213278272" observedRunningTime="2025-12-01 21:51:51.137462788 +0000 UTC m=+1029.627525105" watchObservedRunningTime="2025-12-01 21:51:51.138697019 +0000 UTC m=+1029.628759336" Dec 01 21:51:52 crc kubenswrapper[4857]: I1201 21:51:52.639921 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7f765fc7d6-x797d" Dec 01 21:51:53 crc kubenswrapper[4857]: I1201 21:51:53.636162 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-777b97d487-s7g6x" Dec 01 21:51:57 crc kubenswrapper[4857]: I1201 21:51:57.815959 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:51:57 crc kubenswrapper[4857]: I1201 21:51:57.816482 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:52:02 crc kubenswrapper[4857]: I1201 21:52:02.611021 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-54xtc" Dec 01 21:52:03 crc kubenswrapper[4857]: I1201 21:52:03.293353 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c" Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:03.818554 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7" event={"ID":"dcef9388-5d9b-4172-bb80-ec69b3e98dd0","Type":"ContainerStarted","Data":"48062bef34e07d283e6f8e92d12747dd02b3eb2a5d6087ffaf5a06bb0623a947"} Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:03.819874 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng" event={"ID":"f6e43ef6-1269-40f5-af4d-fe2f30dea319","Type":"ContainerStarted","Data":"47a6d8bb5e8cf8f31591268edcff489c141b087fcc327606fd8c86463c9eace1"} Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:03.820865 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt" event={"ID":"de53c246-6e20-4b67-be68-1c5356fd46d4","Type":"ContainerStarted","Data":"3549daf519a6788d05fd669d26f6d8e8c7ca83ddb5e3cde91d2236d65c10f85a"} Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:03.821876 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk" event={"ID":"80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4","Type":"ContainerStarted","Data":"66a99e2e1ab38462174fc82a08f8476ca281bf1fb9618506bc4cc1dd72f2a58e"} Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:03.823113 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f" event={"ID":"5dd6bfb5-2edc-49b4-8877-85ab5722cdf0","Type":"ContainerStarted","Data":"94d0058d1d570baaa120578633fc5a94dbbdf82542dd82901a588c96d4884cf2"} Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:03.824281 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t" event={"ID":"b13a13cc-0689-480a-bc95-e19aba327f9c","Type":"ContainerStarted","Data":"05c6ddbf0a6febc8eedb198face881655ad9576f1db04e97291a7b91ee42926a"} Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:03.825510 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9" event={"ID":"068a21df-001f-45e6-9946-68f67bad7d43","Type":"ContainerStarted","Data":"00325e2d95280373b8ecd5e2c5d01629083465b2d027fadcc1f205cc32e02a27"} Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:03.826472 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wccvn" event={"ID":"4c8d087e-e00c-4dad-87c7-06ab1017abaf","Type":"ContainerStarted","Data":"72977b3218eab3e65b3ed27801db7e1ef67c3d9204b72b2be4a40583d3d2a919"} Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:03.827532 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb" event={"ID":"92c43db5-8128-4269-9ad4-a263de88f450","Type":"ContainerStarted","Data":"244cf3d9f7a9cdefb2b09a6bd44feab34d5db89ad6e0839f416e809fd4afb500"} Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:04.834687 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt" Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:04.835399 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wccvn" Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:04.835696 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk" Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:04.888062 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng" podStartSLOduration=4.096107744 podStartE2EDuration="58.888013503s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.589342709 +0000 UTC m=+987.079405026" lastFinishedPulling="2025-12-01 21:52:03.381248458 +0000 UTC m=+1041.871310785" observedRunningTime="2025-12-01 21:52:04.878140405 +0000 UTC m=+1043.368202732" watchObservedRunningTime="2025-12-01 21:52:04.888013503 +0000 UTC m=+1043.378075820" Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:04.905093 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt" podStartSLOduration=4.26039891 podStartE2EDuration="58.905069375s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.682583621 +0000 UTC m=+987.172645938" lastFinishedPulling="2025-12-01 21:52:03.327254096 +0000 UTC m=+1041.817316403" observedRunningTime="2025-12-01 21:52:04.90160551 +0000 UTC m=+1043.391667827" watchObservedRunningTime="2025-12-01 21:52:04.905069375 +0000 UTC m=+1043.395131692" Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:04.960951 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7" podStartSLOduration=4.389578072 podStartE2EDuration="58.960926968s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.812477504 +0000 UTC m=+987.302539821" lastFinishedPulling="2025-12-01 21:52:03.3838264 +0000 UTC m=+1041.873888717" observedRunningTime="2025-12-01 21:52:04.951510168 +0000 UTC m=+1043.441572485" watchObservedRunningTime="2025-12-01 21:52:04.960926968 +0000 UTC m=+1043.450989295" Dec 01 21:52:04 crc kubenswrapper[4857]: I1201 21:52:04.981207 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wccvn" podStartSLOduration=4.581931829 podStartE2EDuration="58.981192002s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.927764639 +0000 UTC m=+987.417826956" lastFinishedPulling="2025-12-01 21:52:03.327024802 +0000 UTC m=+1041.817087129" observedRunningTime="2025-12-01 21:52:04.972775948 +0000 UTC m=+1043.462838275" watchObservedRunningTime="2025-12-01 21:52:04.981192002 +0000 UTC m=+1043.471254309" Dec 01 21:52:05 crc kubenswrapper[4857]: I1201 21:52:05.010156 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb" podStartSLOduration=4.307037164 podStartE2EDuration="59.010136874s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.676554093 +0000 UTC m=+987.166616410" lastFinishedPulling="2025-12-01 21:52:03.379653793 +0000 UTC m=+1041.869716120" observedRunningTime="2025-12-01 21:52:05.001838972 +0000 UTC m=+1043.491901299" watchObservedRunningTime="2025-12-01 21:52:05.010136874 +0000 UTC m=+1043.500199191" Dec 01 21:52:05 crc kubenswrapper[4857]: I1201 21:52:05.027604 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk" podStartSLOduration=4.033288223 podStartE2EDuration="59.027588243s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.332946875 +0000 UTC m=+986.823009192" lastFinishedPulling="2025-12-01 21:52:03.327246895 +0000 UTC m=+1041.817309212" observedRunningTime="2025-12-01 21:52:05.02614037 +0000 UTC m=+1043.516202687" watchObservedRunningTime="2025-12-01 21:52:05.027588243 +0000 UTC m=+1043.517650560" Dec 01 21:52:05 crc kubenswrapper[4857]: I1201 21:52:05.048651 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t" podStartSLOduration=4.323479422 podStartE2EDuration="59.048628239s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.648166525 +0000 UTC m=+987.138228842" lastFinishedPulling="2025-12-01 21:52:03.373315352 +0000 UTC m=+1041.863377659" observedRunningTime="2025-12-01 21:52:05.047919288 +0000 UTC m=+1043.537981605" watchObservedRunningTime="2025-12-01 21:52:05.048628239 +0000 UTC m=+1043.538690556" Dec 01 21:52:05 crc kubenswrapper[4857]: I1201 21:52:05.073504 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9" podStartSLOduration=4.492672947 podStartE2EDuration="59.073488016s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.801671579 +0000 UTC m=+987.291733886" lastFinishedPulling="2025-12-01 21:52:03.382486628 +0000 UTC m=+1041.872548955" observedRunningTime="2025-12-01 21:52:05.068246413 +0000 UTC m=+1043.558308730" watchObservedRunningTime="2025-12-01 21:52:05.073488016 +0000 UTC m=+1043.563550333" Dec 01 21:52:05 crc kubenswrapper[4857]: I1201 21:52:05.101075 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f" podStartSLOduration=4.203830999 podStartE2EDuration="59.101057997s" podCreationTimestamp="2025-12-01 21:51:06 +0000 UTC" firstStartedPulling="2025-12-01 21:51:08.48647999 +0000 UTC m=+986.976542307" lastFinishedPulling="2025-12-01 21:52:03.383706988 +0000 UTC m=+1041.873769305" observedRunningTime="2025-12-01 21:52:05.097669813 +0000 UTC m=+1043.587732130" watchObservedRunningTime="2025-12-01 21:52:05.101057997 +0000 UTC m=+1043.591120314" Dec 01 21:52:06 crc kubenswrapper[4857]: I1201 21:52:06.838950 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f" Dec 01 21:52:07 crc kubenswrapper[4857]: I1201 21:52:07.009989 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9" Dec 01 21:52:07 crc kubenswrapper[4857]: I1201 21:52:07.086524 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7" Dec 01 21:52:07 crc kubenswrapper[4857]: I1201 21:52:07.095132 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb" Dec 01 21:52:07 crc kubenswrapper[4857]: I1201 21:52:07.120154 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t" Dec 01 21:52:07 crc kubenswrapper[4857]: I1201 21:52:07.404501 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng" Dec 01 21:52:16 crc kubenswrapper[4857]: I1201 21:52:16.813104 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqfzk" Dec 01 21:52:16 crc kubenswrapper[4857]: I1201 21:52:16.846336 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-j277f" Dec 01 21:52:17 crc kubenswrapper[4857]: I1201 21:52:17.011542 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-zkqw9" Dec 01 21:52:17 crc kubenswrapper[4857]: I1201 21:52:17.090912 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-vtcx7" Dec 01 21:52:17 crc kubenswrapper[4857]: I1201 21:52:17.098918 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b7ksb" Dec 01 21:52:17 crc kubenswrapper[4857]: I1201 21:52:17.121862 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f9d2t" Dec 01 21:52:17 crc kubenswrapper[4857]: I1201 21:52:17.274976 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vbvnt" Dec 01 21:52:17 crc kubenswrapper[4857]: I1201 21:52:17.408910 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-7vmng" Dec 01 21:52:17 crc kubenswrapper[4857]: I1201 21:52:17.594383 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wccvn" Dec 01 21:52:27 crc kubenswrapper[4857]: I1201 21:52:27.806737 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:52:27 crc kubenswrapper[4857]: I1201 21:52:27.807844 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:52:27 crc kubenswrapper[4857]: I1201 21:52:27.807977 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:52:27 crc kubenswrapper[4857]: I1201 21:52:27.809080 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1936180283cd1643268f62d5199bdc5d47710e031991bbd7a33aafc43d3ee862"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 21:52:27 crc kubenswrapper[4857]: I1201 21:52:27.809164 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://1936180283cd1643268f62d5199bdc5d47710e031991bbd7a33aafc43d3ee862" gracePeriod=600 Dec 01 21:52:28 crc kubenswrapper[4857]: I1201 21:52:28.044374 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="1936180283cd1643268f62d5199bdc5d47710e031991bbd7a33aafc43d3ee862" exitCode=0 Dec 01 21:52:28 crc kubenswrapper[4857]: I1201 21:52:28.044416 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"1936180283cd1643268f62d5199bdc5d47710e031991bbd7a33aafc43d3ee862"} Dec 01 21:52:28 crc kubenswrapper[4857]: I1201 21:52:28.044458 4857 scope.go:117] "RemoveContainer" containerID="f890539e4830c07b0e2e556ca8ce83ae288e8de08c435fa380b5373247a6c2d8" Dec 01 21:52:29 crc kubenswrapper[4857]: I1201 21:52:29.061147 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"ae0326e43bb280b472fd2b09c89e53e5301501138efae77ad81d099184e7a849"} Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.500555 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tzbf8"] Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.503382 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.508105 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.516673 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.516946 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-n4rph" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.517176 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.529441 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tzbf8"] Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.569649 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f-config\") pod \"dnsmasq-dns-675f4bcbfc-tzbf8\" (UID: \"6bc5b2d4-4486-4f2e-98fb-f8280ccb019f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.569773 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vp65\" (UniqueName: \"kubernetes.io/projected/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f-kube-api-access-7vp65\") pod \"dnsmasq-dns-675f4bcbfc-tzbf8\" (UID: \"6bc5b2d4-4486-4f2e-98fb-f8280ccb019f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.572672 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-24mkd"] Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.574893 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.586274 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.605921 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-24mkd"] Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.671801 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f-config\") pod \"dnsmasq-dns-675f4bcbfc-tzbf8\" (UID: \"6bc5b2d4-4486-4f2e-98fb-f8280ccb019f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.671902 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vp65\" (UniqueName: \"kubernetes.io/projected/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f-kube-api-access-7vp65\") pod \"dnsmasq-dns-675f4bcbfc-tzbf8\" (UID: \"6bc5b2d4-4486-4f2e-98fb-f8280ccb019f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.673175 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f-config\") pod \"dnsmasq-dns-675f4bcbfc-tzbf8\" (UID: \"6bc5b2d4-4486-4f2e-98fb-f8280ccb019f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.694033 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vp65\" (UniqueName: \"kubernetes.io/projected/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f-kube-api-access-7vp65\") pod \"dnsmasq-dns-675f4bcbfc-tzbf8\" (UID: \"6bc5b2d4-4486-4f2e-98fb-f8280ccb019f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.773268 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57b683aa-6317-44b4-abc3-2381e0e8cf19-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-24mkd\" (UID: \"57b683aa-6317-44b4-abc3-2381e0e8cf19\") " pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.773414 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57b683aa-6317-44b4-abc3-2381e0e8cf19-config\") pod \"dnsmasq-dns-78dd6ddcc-24mkd\" (UID: \"57b683aa-6317-44b4-abc3-2381e0e8cf19\") " pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.773461 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhpf8\" (UniqueName: \"kubernetes.io/projected/57b683aa-6317-44b4-abc3-2381e0e8cf19-kube-api-access-rhpf8\") pod \"dnsmasq-dns-78dd6ddcc-24mkd\" (UID: \"57b683aa-6317-44b4-abc3-2381e0e8cf19\") " pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.842506 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.876308 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57b683aa-6317-44b4-abc3-2381e0e8cf19-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-24mkd\" (UID: \"57b683aa-6317-44b4-abc3-2381e0e8cf19\") " pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.876370 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57b683aa-6317-44b4-abc3-2381e0e8cf19-config\") pod \"dnsmasq-dns-78dd6ddcc-24mkd\" (UID: \"57b683aa-6317-44b4-abc3-2381e0e8cf19\") " pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.876408 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhpf8\" (UniqueName: \"kubernetes.io/projected/57b683aa-6317-44b4-abc3-2381e0e8cf19-kube-api-access-rhpf8\") pod \"dnsmasq-dns-78dd6ddcc-24mkd\" (UID: \"57b683aa-6317-44b4-abc3-2381e0e8cf19\") " pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.877511 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57b683aa-6317-44b4-abc3-2381e0e8cf19-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-24mkd\" (UID: \"57b683aa-6317-44b4-abc3-2381e0e8cf19\") " pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.877827 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57b683aa-6317-44b4-abc3-2381e0e8cf19-config\") pod \"dnsmasq-dns-78dd6ddcc-24mkd\" (UID: \"57b683aa-6317-44b4-abc3-2381e0e8cf19\") " pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.898704 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhpf8\" (UniqueName: \"kubernetes.io/projected/57b683aa-6317-44b4-abc3-2381e0e8cf19-kube-api-access-rhpf8\") pod \"dnsmasq-dns-78dd6ddcc-24mkd\" (UID: \"57b683aa-6317-44b4-abc3-2381e0e8cf19\") " pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:52:33 crc kubenswrapper[4857]: I1201 21:52:33.912119 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:52:34 crc kubenswrapper[4857]: I1201 21:52:34.393542 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tzbf8"] Dec 01 21:52:34 crc kubenswrapper[4857]: I1201 21:52:34.440000 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-24mkd"] Dec 01 21:52:35 crc kubenswrapper[4857]: I1201 21:52:35.141804 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" event={"ID":"57b683aa-6317-44b4-abc3-2381e0e8cf19","Type":"ContainerStarted","Data":"83e7f686e16dad8fcfcf5a57a20d99fbd7893f06b0588609854066a14449aa83"} Dec 01 21:52:35 crc kubenswrapper[4857]: I1201 21:52:35.143910 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" event={"ID":"6bc5b2d4-4486-4f2e-98fb-f8280ccb019f","Type":"ContainerStarted","Data":"ee253911e34f4881440d130512fbe5e69c1c3d89224ddf78dc52768c1124400f"} Dec 01 21:52:36 crc kubenswrapper[4857]: I1201 21:52:36.751682 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tzbf8"] Dec 01 21:52:36 crc kubenswrapper[4857]: I1201 21:52:36.863884 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mw6gs"] Dec 01 21:52:36 crc kubenswrapper[4857]: I1201 21:52:36.865453 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:52:36 crc kubenswrapper[4857]: I1201 21:52:36.895421 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mw6gs"] Dec 01 21:52:36 crc kubenswrapper[4857]: I1201 21:52:36.952464 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fmsw\" (UniqueName: \"kubernetes.io/projected/b0253178-b993-476c-8d26-844ed2851e06-kube-api-access-6fmsw\") pod \"dnsmasq-dns-666b6646f7-mw6gs\" (UID: \"b0253178-b993-476c-8d26-844ed2851e06\") " pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:52:36 crc kubenswrapper[4857]: I1201 21:52:36.952513 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0253178-b993-476c-8d26-844ed2851e06-config\") pod \"dnsmasq-dns-666b6646f7-mw6gs\" (UID: \"b0253178-b993-476c-8d26-844ed2851e06\") " pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:52:36 crc kubenswrapper[4857]: I1201 21:52:36.952570 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0253178-b993-476c-8d26-844ed2851e06-dns-svc\") pod \"dnsmasq-dns-666b6646f7-mw6gs\" (UID: \"b0253178-b993-476c-8d26-844ed2851e06\") " pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.054100 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fmsw\" (UniqueName: \"kubernetes.io/projected/b0253178-b993-476c-8d26-844ed2851e06-kube-api-access-6fmsw\") pod \"dnsmasq-dns-666b6646f7-mw6gs\" (UID: \"b0253178-b993-476c-8d26-844ed2851e06\") " pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.054154 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0253178-b993-476c-8d26-844ed2851e06-config\") pod \"dnsmasq-dns-666b6646f7-mw6gs\" (UID: \"b0253178-b993-476c-8d26-844ed2851e06\") " pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.054224 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0253178-b993-476c-8d26-844ed2851e06-dns-svc\") pod \"dnsmasq-dns-666b6646f7-mw6gs\" (UID: \"b0253178-b993-476c-8d26-844ed2851e06\") " pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.055316 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0253178-b993-476c-8d26-844ed2851e06-dns-svc\") pod \"dnsmasq-dns-666b6646f7-mw6gs\" (UID: \"b0253178-b993-476c-8d26-844ed2851e06\") " pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.055493 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0253178-b993-476c-8d26-844ed2851e06-config\") pod \"dnsmasq-dns-666b6646f7-mw6gs\" (UID: \"b0253178-b993-476c-8d26-844ed2851e06\") " pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.137367 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fmsw\" (UniqueName: \"kubernetes.io/projected/b0253178-b993-476c-8d26-844ed2851e06-kube-api-access-6fmsw\") pod \"dnsmasq-dns-666b6646f7-mw6gs\" (UID: \"b0253178-b993-476c-8d26-844ed2851e06\") " pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.202382 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.319161 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-24mkd"] Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.395286 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-l7k8n"] Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.396936 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.409098 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-l7k8n"] Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.479503 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69b986d7-590f-4fa1-b2ab-2bef95fe745e-config\") pod \"dnsmasq-dns-57d769cc4f-l7k8n\" (UID: \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\") " pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.480061 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxtst\" (UniqueName: \"kubernetes.io/projected/69b986d7-590f-4fa1-b2ab-2bef95fe745e-kube-api-access-fxtst\") pod \"dnsmasq-dns-57d769cc4f-l7k8n\" (UID: \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\") " pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.480101 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69b986d7-590f-4fa1-b2ab-2bef95fe745e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-l7k8n\" (UID: \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\") " pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.873358 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxtst\" (UniqueName: \"kubernetes.io/projected/69b986d7-590f-4fa1-b2ab-2bef95fe745e-kube-api-access-fxtst\") pod \"dnsmasq-dns-57d769cc4f-l7k8n\" (UID: \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\") " pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.873430 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69b986d7-590f-4fa1-b2ab-2bef95fe745e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-l7k8n\" (UID: \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\") " pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.873479 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69b986d7-590f-4fa1-b2ab-2bef95fe745e-config\") pod \"dnsmasq-dns-57d769cc4f-l7k8n\" (UID: \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\") " pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.874594 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69b986d7-590f-4fa1-b2ab-2bef95fe745e-config\") pod \"dnsmasq-dns-57d769cc4f-l7k8n\" (UID: \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\") " pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.889468 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69b986d7-590f-4fa1-b2ab-2bef95fe745e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-l7k8n\" (UID: \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\") " pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:52:37 crc kubenswrapper[4857]: I1201 21:52:37.920957 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxtst\" (UniqueName: \"kubernetes.io/projected/69b986d7-590f-4fa1-b2ab-2bef95fe745e-kube-api-access-fxtst\") pod \"dnsmasq-dns-57d769cc4f-l7k8n\" (UID: \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\") " pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.024822 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.054449 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.056555 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.063238 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.067545 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.075464 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.075809 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.075823 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zgfph" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.075915 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.075939 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.116773 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.185821 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-config-data\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.185971 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.186530 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.186600 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdqld\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-kube-api-access-qdqld\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.186945 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.193201 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.193284 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.193327 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.193368 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.193410 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/235e0415-1e77-480d-b3da-157ed0252d4f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.193456 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/235e0415-1e77-480d-b3da-157ed0252d4f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.295863 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.296443 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.296481 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.296510 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/235e0415-1e77-480d-b3da-157ed0252d4f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.296542 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/235e0415-1e77-480d-b3da-157ed0252d4f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.296570 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-config-data\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.296600 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.296643 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.296661 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdqld\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-kube-api-access-qdqld\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.296728 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.296809 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.301637 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.302498 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.304093 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.304434 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-config-data\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.304760 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.304778 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.317604 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/235e0415-1e77-480d-b3da-157ed0252d4f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.319693 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.319779 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.322733 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/235e0415-1e77-480d-b3da-157ed0252d4f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.359742 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdqld\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-kube-api-access-qdqld\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.370423 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: W1201 21:52:38.499873 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0253178_b993_476c_8d26_844ed2851e06.slice/crio-2fe1fe74fc90c941a3eabe087b2f2bb2b88fa7719d9a963b78d289ebb4d69302 WatchSource:0}: Error finding container 2fe1fe74fc90c941a3eabe087b2f2bb2b88fa7719d9a963b78d289ebb4d69302: Status 404 returned error can't find the container with id 2fe1fe74fc90c941a3eabe087b2f2bb2b88fa7719d9a963b78d289ebb4d69302 Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.506963 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mw6gs"] Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.771744 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 21:52:38 crc kubenswrapper[4857]: I1201 21:52:38.851802 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-l7k8n"] Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.005585 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.007320 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.013981 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.023643 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.023958 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.024147 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.024286 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.024508 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.024673 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-57hqb" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.025550 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.200928 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twjb6\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-kube-api-access-twjb6\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.201026 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.201226 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.201289 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.201409 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.201442 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.201531 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84a09f7d-b0a7-4055-93e1-83d11edb6467-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.206105 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.206182 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.206220 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.206370 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84a09f7d-b0a7-4055-93e1-83d11edb6467-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.218130 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" event={"ID":"b0253178-b993-476c-8d26-844ed2851e06","Type":"ContainerStarted","Data":"2fe1fe74fc90c941a3eabe087b2f2bb2b88fa7719d9a963b78d289ebb4d69302"} Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.228249 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" event={"ID":"69b986d7-590f-4fa1-b2ab-2bef95fe745e","Type":"ContainerStarted","Data":"e421e0aff90b3152383b2d4260fcd344e57348be1a01b7b369c892f9f368a478"} Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.331341 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.331381 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.331401 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.331438 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84a09f7d-b0a7-4055-93e1-83d11edb6467-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.331466 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.331482 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twjb6\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-kube-api-access-twjb6\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.331518 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.331541 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.331566 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.331580 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.331600 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84a09f7d-b0a7-4055-93e1-83d11edb6467-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.333349 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.334213 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.334498 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.335501 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.335663 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.339604 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84a09f7d-b0a7-4055-93e1-83d11edb6467-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.340721 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.344084 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.366200 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84a09f7d-b0a7-4055-93e1-83d11edb6467-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.367848 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.368389 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.415017 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twjb6\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-kube-api-access-twjb6\") pod \"rabbitmq-cell1-server-0\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.541492 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 21:52:39 crc kubenswrapper[4857]: I1201 21:52:39.679344 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.152331 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.154349 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.157119 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.158599 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-mc6c2" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.158770 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.159465 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.161148 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.175514 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.259945 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"235e0415-1e77-480d-b3da-157ed0252d4f","Type":"ContainerStarted","Data":"465ef4ddbbe189e930cce6323f28c5f8065568a5db7f8cb04fb69271dd658c94"} Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.270471 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3755be2-a089-4259-af43-6e0fbedef24b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.270535 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b3755be2-a089-4259-af43-6e0fbedef24b-config-data-default\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.270562 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3755be2-a089-4259-af43-6e0fbedef24b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.270579 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25jnx\" (UniqueName: \"kubernetes.io/projected/b3755be2-a089-4259-af43-6e0fbedef24b-kube-api-access-25jnx\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.270618 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b3755be2-a089-4259-af43-6e0fbedef24b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.270663 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.271015 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b3755be2-a089-4259-af43-6e0fbedef24b-kolla-config\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.271094 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3755be2-a089-4259-af43-6e0fbedef24b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.377557 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3755be2-a089-4259-af43-6e0fbedef24b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.377648 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3755be2-a089-4259-af43-6e0fbedef24b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.377706 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b3755be2-a089-4259-af43-6e0fbedef24b-config-data-default\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.377730 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3755be2-a089-4259-af43-6e0fbedef24b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.377769 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25jnx\" (UniqueName: \"kubernetes.io/projected/b3755be2-a089-4259-af43-6e0fbedef24b-kube-api-access-25jnx\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.377822 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b3755be2-a089-4259-af43-6e0fbedef24b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.377872 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.377901 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b3755be2-a089-4259-af43-6e0fbedef24b-kolla-config\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.379232 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b3755be2-a089-4259-af43-6e0fbedef24b-kolla-config\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.388618 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b3755be2-a089-4259-af43-6e0fbedef24b-config-data-default\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.389441 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b3755be2-a089-4259-af43-6e0fbedef24b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.389653 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.390004 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3755be2-a089-4259-af43-6e0fbedef24b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.393636 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3755be2-a089-4259-af43-6e0fbedef24b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.416361 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25jnx\" (UniqueName: \"kubernetes.io/projected/b3755be2-a089-4259-af43-6e0fbedef24b-kube-api-access-25jnx\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.417097 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3755be2-a089-4259-af43-6e0fbedef24b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.444468 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.454839 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"b3755be2-a089-4259-af43-6e0fbedef24b\") " pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.525140 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.913891 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 21:52:40 crc kubenswrapper[4857]: I1201 21:52:40.976639 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.026203 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.030300 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.039422 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.039711 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-5jdpj" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.039848 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.055366 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.272863 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/be0a242c-3c02-4d1e-910b-21e74e7c4e43-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.281121 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be0a242c-3c02-4d1e-910b-21e74e7c4e43-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.281185 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/be0a242c-3c02-4d1e-910b-21e74e7c4e43-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.287473 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.287667 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0a242c-3c02-4d1e-910b-21e74e7c4e43-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.287714 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/be0a242c-3c02-4d1e-910b-21e74e7c4e43-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.287731 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5scwq\" (UniqueName: \"kubernetes.io/projected/be0a242c-3c02-4d1e-910b-21e74e7c4e43-kube-api-access-5scwq\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.287768 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/be0a242c-3c02-4d1e-910b-21e74e7c4e43-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.302288 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b3755be2-a089-4259-af43-6e0fbedef24b","Type":"ContainerStarted","Data":"fb24ffcad4c31897c6d62e5f2123740e4ad0219d782e94c889c7172d1855ab72"} Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.328771 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"84a09f7d-b0a7-4055-93e1-83d11edb6467","Type":"ContainerStarted","Data":"dfe281ab23930f2c45d9b50e610b8b139302b96aff16fe846f5a7a987ad525be"} Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.355584 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.363317 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.365085 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.373734 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.373999 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-lxlft" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.374652 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399009 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399129 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0a242c-3c02-4d1e-910b-21e74e7c4e43-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399177 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5scwq\" (UniqueName: \"kubernetes.io/projected/be0a242c-3c02-4d1e-910b-21e74e7c4e43-kube-api-access-5scwq\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399196 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/be0a242c-3c02-4d1e-910b-21e74e7c4e43-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399217 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/be0a242c-3c02-4d1e-910b-21e74e7c4e43-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399250 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a290cd2-c321-4cd5-b168-4866c03b135b-config-data\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399270 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scv45\" (UniqueName: \"kubernetes.io/projected/6a290cd2-c321-4cd5-b168-4866c03b135b-kube-api-access-scv45\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399336 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a290cd2-c321-4cd5-b168-4866c03b135b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399354 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/be0a242c-3c02-4d1e-910b-21e74e7c4e43-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399437 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be0a242c-3c02-4d1e-910b-21e74e7c4e43-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399463 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6a290cd2-c321-4cd5-b168-4866c03b135b-kolla-config\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399503 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/be0a242c-3c02-4d1e-910b-21e74e7c4e43-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399531 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a290cd2-c321-4cd5-b168-4866c03b135b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.399928 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.400331 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/be0a242c-3c02-4d1e-910b-21e74e7c4e43-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.401693 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/be0a242c-3c02-4d1e-910b-21e74e7c4e43-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.402117 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/be0a242c-3c02-4d1e-910b-21e74e7c4e43-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.403190 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be0a242c-3c02-4d1e-910b-21e74e7c4e43-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.421847 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/be0a242c-3c02-4d1e-910b-21e74e7c4e43-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.425637 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0a242c-3c02-4d1e-910b-21e74e7c4e43-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.462152 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5scwq\" (UniqueName: \"kubernetes.io/projected/be0a242c-3c02-4d1e-910b-21e74e7c4e43-kube-api-access-5scwq\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.483063 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"be0a242c-3c02-4d1e-910b-21e74e7c4e43\") " pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.507099 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a290cd2-c321-4cd5-b168-4866c03b135b-config-data\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.507144 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scv45\" (UniqueName: \"kubernetes.io/projected/6a290cd2-c321-4cd5-b168-4866c03b135b-kube-api-access-scv45\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.507190 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a290cd2-c321-4cd5-b168-4866c03b135b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.507237 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6a290cd2-c321-4cd5-b168-4866c03b135b-kolla-config\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.507260 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a290cd2-c321-4cd5-b168-4866c03b135b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.513073 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a290cd2-c321-4cd5-b168-4866c03b135b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.516666 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a290cd2-c321-4cd5-b168-4866c03b135b-config-data\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.517143 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6a290cd2-c321-4cd5-b168-4866c03b135b-kolla-config\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.526449 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a290cd2-c321-4cd5-b168-4866c03b135b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.597272 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scv45\" (UniqueName: \"kubernetes.io/projected/6a290cd2-c321-4cd5-b168-4866c03b135b-kube-api-access-scv45\") pod \"memcached-0\" (UID: \"6a290cd2-c321-4cd5-b168-4866c03b135b\") " pod="openstack/memcached-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.673434 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 21:52:41 crc kubenswrapper[4857]: I1201 21:52:41.708775 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 21:52:42 crc kubenswrapper[4857]: I1201 21:52:42.688130 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 21:52:42 crc kubenswrapper[4857]: I1201 21:52:42.923853 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 21:52:43 crc kubenswrapper[4857]: I1201 21:52:43.382772 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"be0a242c-3c02-4d1e-910b-21e74e7c4e43","Type":"ContainerStarted","Data":"0da3916f5bee588d1d54d919456a32e932039ab98ab961a8970d738927c5b93c"} Dec 01 21:52:43 crc kubenswrapper[4857]: I1201 21:52:43.385484 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6a290cd2-c321-4cd5-b168-4866c03b135b","Type":"ContainerStarted","Data":"868e1019fd711cf1d70b23ba69452586e12c2ed094069d1f7a1624de7acdb46f"} Dec 01 21:52:43 crc kubenswrapper[4857]: I1201 21:52:43.666525 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 21:52:43 crc kubenswrapper[4857]: I1201 21:52:43.667483 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 21:52:43 crc kubenswrapper[4857]: I1201 21:52:43.693679 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-7wsdv" Dec 01 21:52:43 crc kubenswrapper[4857]: I1201 21:52:43.694522 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 21:52:43 crc kubenswrapper[4857]: I1201 21:52:43.758620 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrtbd\" (UniqueName: \"kubernetes.io/projected/b2f9d1d0-36e7-44f0-9c03-3817bafe85ce-kube-api-access-wrtbd\") pod \"kube-state-metrics-0\" (UID: \"b2f9d1d0-36e7-44f0-9c03-3817bafe85ce\") " pod="openstack/kube-state-metrics-0" Dec 01 21:52:43 crc kubenswrapper[4857]: I1201 21:52:43.878813 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrtbd\" (UniqueName: \"kubernetes.io/projected/b2f9d1d0-36e7-44f0-9c03-3817bafe85ce-kube-api-access-wrtbd\") pod \"kube-state-metrics-0\" (UID: \"b2f9d1d0-36e7-44f0-9c03-3817bafe85ce\") " pod="openstack/kube-state-metrics-0" Dec 01 21:52:43 crc kubenswrapper[4857]: I1201 21:52:43.925187 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrtbd\" (UniqueName: \"kubernetes.io/projected/b2f9d1d0-36e7-44f0-9c03-3817bafe85ce-kube-api-access-wrtbd\") pod \"kube-state-metrics-0\" (UID: \"b2f9d1d0-36e7-44f0-9c03-3817bafe85ce\") " pod="openstack/kube-state-metrics-0" Dec 01 21:52:44 crc kubenswrapper[4857]: I1201 21:52:44.015350 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 21:52:44 crc kubenswrapper[4857]: I1201 21:52:44.961290 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 21:52:45 crc kubenswrapper[4857]: W1201 21:52:45.010956 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2f9d1d0_36e7_44f0_9c03_3817bafe85ce.slice/crio-95e7ecc583ad65a7a80d8d391a951aff950101e3eae5126402f1d5be3cdbc2ac WatchSource:0}: Error finding container 95e7ecc583ad65a7a80d8d391a951aff950101e3eae5126402f1d5be3cdbc2ac: Status 404 returned error can't find the container with id 95e7ecc583ad65a7a80d8d391a951aff950101e3eae5126402f1d5be3cdbc2ac Dec 01 21:52:45 crc kubenswrapper[4857]: I1201 21:52:45.545358 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b2f9d1d0-36e7-44f0-9c03-3817bafe85ce","Type":"ContainerStarted","Data":"95e7ecc583ad65a7a80d8d391a951aff950101e3eae5126402f1d5be3cdbc2ac"} Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.006828 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-nhfhn"] Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.008454 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.011050 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.011227 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-l97m9" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.012320 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.019554 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nhfhn"] Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.027746 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-gwbmr"] Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.029390 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.033146 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gwbmr"] Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.111629 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-var-lib\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.111686 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/32213ad3-4287-4968-9d42-bfeb2ff558d1-ovn-controller-tls-certs\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.111737 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-var-run\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.111877 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/32213ad3-4287-4968-9d42-bfeb2ff558d1-var-run\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.112003 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-scripts\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.112087 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32213ad3-4287-4968-9d42-bfeb2ff558d1-combined-ca-bundle\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.116138 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9x8h\" (UniqueName: \"kubernetes.io/projected/32213ad3-4287-4968-9d42-bfeb2ff558d1-kube-api-access-f9x8h\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.116199 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32213ad3-4287-4968-9d42-bfeb2ff558d1-scripts\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.116255 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr5zt\" (UniqueName: \"kubernetes.io/projected/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-kube-api-access-sr5zt\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.116426 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-var-log\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.116487 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-etc-ovs\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.116530 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/32213ad3-4287-4968-9d42-bfeb2ff558d1-var-run-ovn\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.116573 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/32213ad3-4287-4968-9d42-bfeb2ff558d1-var-log-ovn\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218145 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr5zt\" (UniqueName: \"kubernetes.io/projected/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-kube-api-access-sr5zt\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218249 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-var-log\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218305 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-etc-ovs\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218335 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/32213ad3-4287-4968-9d42-bfeb2ff558d1-var-run-ovn\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218359 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/32213ad3-4287-4968-9d42-bfeb2ff558d1-var-log-ovn\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218384 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-var-lib\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218407 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/32213ad3-4287-4968-9d42-bfeb2ff558d1-ovn-controller-tls-certs\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218442 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-var-run\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218477 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/32213ad3-4287-4968-9d42-bfeb2ff558d1-var-run\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218509 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-scripts\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218560 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32213ad3-4287-4968-9d42-bfeb2ff558d1-combined-ca-bundle\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218919 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/32213ad3-4287-4968-9d42-bfeb2ff558d1-var-run-ovn\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.219086 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-etc-ovs\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.219102 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-var-log\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.219197 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-var-run\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.219277 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/32213ad3-4287-4968-9d42-bfeb2ff558d1-var-log-ovn\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.219401 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-var-lib\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.218588 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9x8h\" (UniqueName: \"kubernetes.io/projected/32213ad3-4287-4968-9d42-bfeb2ff558d1-kube-api-access-f9x8h\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.219768 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32213ad3-4287-4968-9d42-bfeb2ff558d1-scripts\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.221308 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/32213ad3-4287-4968-9d42-bfeb2ff558d1-var-run\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.222362 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-scripts\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.223727 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32213ad3-4287-4968-9d42-bfeb2ff558d1-scripts\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.236920 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/32213ad3-4287-4968-9d42-bfeb2ff558d1-ovn-controller-tls-certs\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.253606 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9x8h\" (UniqueName: \"kubernetes.io/projected/32213ad3-4287-4968-9d42-bfeb2ff558d1-kube-api-access-f9x8h\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.297148 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr5zt\" (UniqueName: \"kubernetes.io/projected/5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f-kube-api-access-sr5zt\") pod \"ovn-controller-ovs-gwbmr\" (UID: \"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f\") " pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.331587 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32213ad3-4287-4968-9d42-bfeb2ff558d1-combined-ca-bundle\") pod \"ovn-controller-nhfhn\" (UID: \"32213ad3-4287-4968-9d42-bfeb2ff558d1\") " pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.419992 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nhfhn" Dec 01 21:52:47 crc kubenswrapper[4857]: I1201 21:52:47.553328 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.243902 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.245846 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.247873 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.247900 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.249129 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.249977 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-jgcjv" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.251104 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.252196 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.352449 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ztk4\" (UniqueName: \"kubernetes.io/projected/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-kube-api-access-5ztk4\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.352508 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.352551 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.352596 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.352627 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.352677 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.352704 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.352748 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-config\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.456975 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ztk4\" (UniqueName: \"kubernetes.io/projected/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-kube-api-access-5ztk4\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.457029 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.457077 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.457109 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.457140 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.457172 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.457198 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.457229 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-config\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.458206 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.460233 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-config\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.460412 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.460709 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.476353 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.476942 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.479315 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.480613 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ztk4\" (UniqueName: \"kubernetes.io/projected/ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3-kube-api-access-5ztk4\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.517828 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3\") " pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:49 crc kubenswrapper[4857]: I1201 21:52:49.590528 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.784666 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.789992 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.796008 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.800570 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-ptqfm" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.800614 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.800832 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.804075 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.958502 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b726f947-1deb-4268-ae76-900394324ecf-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.958563 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b726f947-1deb-4268-ae76-900394324ecf-config\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.958597 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.958619 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b726f947-1deb-4268-ae76-900394324ecf-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.958638 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbn4w\" (UniqueName: \"kubernetes.io/projected/b726f947-1deb-4268-ae76-900394324ecf-kube-api-access-sbn4w\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.958713 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b726f947-1deb-4268-ae76-900394324ecf-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.958744 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b726f947-1deb-4268-ae76-900394324ecf-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:50 crc kubenswrapper[4857]: I1201 21:52:50.958766 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b726f947-1deb-4268-ae76-900394324ecf-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.060760 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b726f947-1deb-4268-ae76-900394324ecf-config\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.060812 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.060840 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b726f947-1deb-4268-ae76-900394324ecf-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.060861 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbn4w\" (UniqueName: \"kubernetes.io/projected/b726f947-1deb-4268-ae76-900394324ecf-kube-api-access-sbn4w\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.060908 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b726f947-1deb-4268-ae76-900394324ecf-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.060939 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b726f947-1deb-4268-ae76-900394324ecf-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.060966 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b726f947-1deb-4268-ae76-900394324ecf-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.061012 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b726f947-1deb-4268-ae76-900394324ecf-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.061322 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.062267 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b726f947-1deb-4268-ae76-900394324ecf-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.063736 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b726f947-1deb-4268-ae76-900394324ecf-config\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.065818 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b726f947-1deb-4268-ae76-900394324ecf-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.071797 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b726f947-1deb-4268-ae76-900394324ecf-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.071882 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b726f947-1deb-4268-ae76-900394324ecf-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.072101 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b726f947-1deb-4268-ae76-900394324ecf-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.082038 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbn4w\" (UniqueName: \"kubernetes.io/projected/b726f947-1deb-4268-ae76-900394324ecf-kube-api-access-sbn4w\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.096125 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b726f947-1deb-4268-ae76-900394324ecf\") " pod="openstack/ovsdbserver-nb-0" Dec 01 21:52:51 crc kubenswrapper[4857]: I1201 21:52:51.173641 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 21:53:05 crc kubenswrapper[4857]: E1201 21:53:05.181564 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 01 21:53:05 crc kubenswrapper[4857]: E1201 21:53:05.182215 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5scwq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(be0a242c-3c02-4d1e-910b-21e74e7c4e43): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:53:05 crc kubenswrapper[4857]: E1201 21:53:05.183698 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="be0a242c-3c02-4d1e-910b-21e74e7c4e43" Dec 01 21:53:05 crc kubenswrapper[4857]: E1201 21:53:05.856774 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="be0a242c-3c02-4d1e-910b-21e74e7c4e43" Dec 01 21:53:06 crc kubenswrapper[4857]: E1201 21:53:06.414584 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Dec 01 21:53:06 crc kubenswrapper[4857]: E1201 21:53:06.414794 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n68h66h75h9fh686h68fh644h556h7ch59dhffh6fh5bbh68fh578h75h67bhcbh54ch5bbh59bh67dh679h5ch5cdhb7h55dh689h57dh9bh544hf9q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-scv45,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(6a290cd2-c321-4cd5-b168-4866c03b135b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:53:06 crc kubenswrapper[4857]: E1201 21:53:06.415959 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="6a290cd2-c321-4cd5-b168-4866c03b135b" Dec 01 21:53:06 crc kubenswrapper[4857]: E1201 21:53:06.483283 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 01 21:53:06 crc kubenswrapper[4857]: E1201 21:53:06.483441 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-25jnx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(b3755be2-a089-4259-af43-6e0fbedef24b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:53:06 crc kubenswrapper[4857]: E1201 21:53:06.484836 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="b3755be2-a089-4259-af43-6e0fbedef24b" Dec 01 21:53:06 crc kubenswrapper[4857]: E1201 21:53:06.863422 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="b3755be2-a089-4259-af43-6e0fbedef24b" Dec 01 21:53:06 crc kubenswrapper[4857]: E1201 21:53:06.863544 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="6a290cd2-c321-4cd5-b168-4866c03b135b" Dec 01 21:53:11 crc kubenswrapper[4857]: E1201 21:53:11.898765 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 21:53:11 crc kubenswrapper[4857]: E1201 21:53:11.899483 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7vp65,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-tzbf8_openstack(6bc5b2d4-4486-4f2e-98fb-f8280ccb019f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:53:11 crc kubenswrapper[4857]: E1201 21:53:11.900729 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" podUID="6bc5b2d4-4486-4f2e-98fb-f8280ccb019f" Dec 01 21:53:11 crc kubenswrapper[4857]: E1201 21:53:11.943021 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 21:53:11 crc kubenswrapper[4857]: E1201 21:53:11.944083 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rhpf8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-24mkd_openstack(57b683aa-6317-44b4-abc3-2381e0e8cf19): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:53:11 crc kubenswrapper[4857]: E1201 21:53:11.945212 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" podUID="57b683aa-6317-44b4-abc3-2381e0e8cf19" Dec 01 21:53:11 crc kubenswrapper[4857]: E1201 21:53:11.988215 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 21:53:11 crc kubenswrapper[4857]: E1201 21:53:11.988375 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6fmsw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-mw6gs_openstack(b0253178-b993-476c-8d26-844ed2851e06): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:53:11 crc kubenswrapper[4857]: E1201 21:53:11.990168 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" podUID="b0253178-b993-476c-8d26-844ed2851e06" Dec 01 21:53:12 crc kubenswrapper[4857]: E1201 21:53:12.010869 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 21:53:12 crc kubenswrapper[4857]: E1201 21:53:12.011150 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fxtst,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-l7k8n_openstack(69b986d7-590f-4fa1-b2ab-2bef95fe745e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:53:12 crc kubenswrapper[4857]: E1201 21:53:12.012573 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" podUID="69b986d7-590f-4fa1-b2ab-2bef95fe745e" Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.301126 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.440137 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f-config\") pod \"6bc5b2d4-4486-4f2e-98fb-f8280ccb019f\" (UID: \"6bc5b2d4-4486-4f2e-98fb-f8280ccb019f\") " Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.440341 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vp65\" (UniqueName: \"kubernetes.io/projected/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f-kube-api-access-7vp65\") pod \"6bc5b2d4-4486-4f2e-98fb-f8280ccb019f\" (UID: \"6bc5b2d4-4486-4f2e-98fb-f8280ccb019f\") " Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.441180 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f-config" (OuterVolumeSpecName: "config") pod "6bc5b2d4-4486-4f2e-98fb-f8280ccb019f" (UID: "6bc5b2d4-4486-4f2e-98fb-f8280ccb019f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.446585 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f-kube-api-access-7vp65" (OuterVolumeSpecName: "kube-api-access-7vp65") pod "6bc5b2d4-4486-4f2e-98fb-f8280ccb019f" (UID: "6bc5b2d4-4486-4f2e-98fb-f8280ccb019f"). InnerVolumeSpecName "kube-api-access-7vp65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.543334 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.543373 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vp65\" (UniqueName: \"kubernetes.io/projected/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f-kube-api-access-7vp65\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.682450 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.699120 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nhfhn"] Dec 01 21:53:12 crc kubenswrapper[4857]: W1201 21:53:12.761088 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb726f947_1deb_4268_ae76_900394324ecf.slice/crio-91f9e245ce736fbbc250228a04a20b2adc9431b980071f6f76b0fa68f39221c9 WatchSource:0}: Error finding container 91f9e245ce736fbbc250228a04a20b2adc9431b980071f6f76b0fa68f39221c9: Status 404 returned error can't find the container with id 91f9e245ce736fbbc250228a04a20b2adc9431b980071f6f76b0fa68f39221c9 Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.783689 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gwbmr"] Dec 01 21:53:12 crc kubenswrapper[4857]: W1201 21:53:12.861159 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32213ad3_4287_4968_9d42_bfeb2ff558d1.slice/crio-616c88a352b41cec953614f2fbe1d01a93d96352bc325c774d56e964b5f116de WatchSource:0}: Error finding container 616c88a352b41cec953614f2fbe1d01a93d96352bc325c774d56e964b5f116de: Status 404 returned error can't find the container with id 616c88a352b41cec953614f2fbe1d01a93d96352bc325c774d56e964b5f116de Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.936300 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nhfhn" event={"ID":"32213ad3-4287-4968-9d42-bfeb2ff558d1","Type":"ContainerStarted","Data":"616c88a352b41cec953614f2fbe1d01a93d96352bc325c774d56e964b5f116de"} Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.937677 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b726f947-1deb-4268-ae76-900394324ecf","Type":"ContainerStarted","Data":"91f9e245ce736fbbc250228a04a20b2adc9431b980071f6f76b0fa68f39221c9"} Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.938688 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" event={"ID":"6bc5b2d4-4486-4f2e-98fb-f8280ccb019f","Type":"ContainerDied","Data":"ee253911e34f4881440d130512fbe5e69c1c3d89224ddf78dc52768c1124400f"} Dec 01 21:53:12 crc kubenswrapper[4857]: I1201 21:53:12.938755 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tzbf8" Dec 01 21:53:12 crc kubenswrapper[4857]: E1201 21:53:12.940933 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" podUID="69b986d7-590f-4fa1-b2ab-2bef95fe745e" Dec 01 21:53:12 crc kubenswrapper[4857]: E1201 21:53:12.941848 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" podUID="b0253178-b993-476c-8d26-844ed2851e06" Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.182679 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tzbf8"] Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.193499 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tzbf8"] Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.288692 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 21:53:13 crc kubenswrapper[4857]: W1201 21:53:13.361972 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a6a8c4c_bbe8_4808_a83a_9f00bcf1919f.slice/crio-d1200b1db4c7b9038d28e837c1cb21de041049be784c50c5ad0c1002c15563e1 WatchSource:0}: Error finding container d1200b1db4c7b9038d28e837c1cb21de041049be784c50c5ad0c1002c15563e1: Status 404 returned error can't find the container with id d1200b1db4c7b9038d28e837c1cb21de041049be784c50c5ad0c1002c15563e1 Dec 01 21:53:13 crc kubenswrapper[4857]: W1201 21:53:13.463875 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecd20b97_c60f_48c4_b2ab_a11ff8fa38f3.slice/crio-62555188bb50576b2ace16917022afc11d5562e9377e5af8330074b7aab87f8b WatchSource:0}: Error finding container 62555188bb50576b2ace16917022afc11d5562e9377e5af8330074b7aab87f8b: Status 404 returned error can't find the container with id 62555188bb50576b2ace16917022afc11d5562e9377e5af8330074b7aab87f8b Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.525396 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.663774 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhpf8\" (UniqueName: \"kubernetes.io/projected/57b683aa-6317-44b4-abc3-2381e0e8cf19-kube-api-access-rhpf8\") pod \"57b683aa-6317-44b4-abc3-2381e0e8cf19\" (UID: \"57b683aa-6317-44b4-abc3-2381e0e8cf19\") " Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.663857 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57b683aa-6317-44b4-abc3-2381e0e8cf19-dns-svc\") pod \"57b683aa-6317-44b4-abc3-2381e0e8cf19\" (UID: \"57b683aa-6317-44b4-abc3-2381e0e8cf19\") " Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.663896 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57b683aa-6317-44b4-abc3-2381e0e8cf19-config\") pod \"57b683aa-6317-44b4-abc3-2381e0e8cf19\" (UID: \"57b683aa-6317-44b4-abc3-2381e0e8cf19\") " Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.665245 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57b683aa-6317-44b4-abc3-2381e0e8cf19-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "57b683aa-6317-44b4-abc3-2381e0e8cf19" (UID: "57b683aa-6317-44b4-abc3-2381e0e8cf19"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.665268 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57b683aa-6317-44b4-abc3-2381e0e8cf19-config" (OuterVolumeSpecName: "config") pod "57b683aa-6317-44b4-abc3-2381e0e8cf19" (UID: "57b683aa-6317-44b4-abc3-2381e0e8cf19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.670638 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b683aa-6317-44b4-abc3-2381e0e8cf19-kube-api-access-rhpf8" (OuterVolumeSpecName: "kube-api-access-rhpf8") pod "57b683aa-6317-44b4-abc3-2381e0e8cf19" (UID: "57b683aa-6317-44b4-abc3-2381e0e8cf19"). InnerVolumeSpecName "kube-api-access-rhpf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.765674 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhpf8\" (UniqueName: \"kubernetes.io/projected/57b683aa-6317-44b4-abc3-2381e0e8cf19-kube-api-access-rhpf8\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.765717 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57b683aa-6317-44b4-abc3-2381e0e8cf19-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.765726 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57b683aa-6317-44b4-abc3-2381e0e8cf19-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.862279 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bc5b2d4-4486-4f2e-98fb-f8280ccb019f" path="/var/lib/kubelet/pods/6bc5b2d4-4486-4f2e-98fb-f8280ccb019f/volumes" Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.956121 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.956134 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-24mkd" event={"ID":"57b683aa-6317-44b4-abc3-2381e0e8cf19","Type":"ContainerDied","Data":"83e7f686e16dad8fcfcf5a57a20d99fbd7893f06b0588609854066a14449aa83"} Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.959874 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gwbmr" event={"ID":"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f","Type":"ContainerStarted","Data":"d1200b1db4c7b9038d28e837c1cb21de041049be784c50c5ad0c1002c15563e1"} Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.961645 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"84a09f7d-b0a7-4055-93e1-83d11edb6467","Type":"ContainerStarted","Data":"8e89121786dac33726d523e59f2c80de2165f73812834f8643c69bd32e7e2d1c"} Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.962629 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3","Type":"ContainerStarted","Data":"62555188bb50576b2ace16917022afc11d5562e9377e5af8330074b7aab87f8b"} Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.964432 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"235e0415-1e77-480d-b3da-157ed0252d4f","Type":"ContainerStarted","Data":"868e89c708562210a5a6e915928b232c0e7618e25b2d87a0152f36a0bcbc14ad"} Dec 01 21:53:13 crc kubenswrapper[4857]: I1201 21:53:13.999226 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-24mkd"] Dec 01 21:53:14 crc kubenswrapper[4857]: I1201 21:53:14.004693 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-24mkd"] Dec 01 21:53:15 crc kubenswrapper[4857]: I1201 21:53:15.845605 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57b683aa-6317-44b4-abc3-2381e0e8cf19" path="/var/lib/kubelet/pods/57b683aa-6317-44b4-abc3-2381e0e8cf19/volumes" Dec 01 21:53:17 crc kubenswrapper[4857]: I1201 21:53:17.998233 4857 generic.go:334] "Generic (PLEG): container finished" podID="5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f" containerID="7f13bc532ae6d68a62ed73764d852e20af171fd307e95f456826579e8e5d03b6" exitCode=0 Dec 01 21:53:18 crc kubenswrapper[4857]: I1201 21:53:17.998468 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gwbmr" event={"ID":"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f","Type":"ContainerDied","Data":"7f13bc532ae6d68a62ed73764d852e20af171fd307e95f456826579e8e5d03b6"} Dec 01 21:53:18 crc kubenswrapper[4857]: I1201 21:53:18.003777 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b2f9d1d0-36e7-44f0-9c03-3817bafe85ce","Type":"ContainerStarted","Data":"79728cdc0cb978317ab106a8a0a8d0a3b9c48a6a8f65af21ceb348ef5560e712"} Dec 01 21:53:18 crc kubenswrapper[4857]: I1201 21:53:18.003973 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 21:53:18 crc kubenswrapper[4857]: I1201 21:53:18.006394 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b726f947-1deb-4268-ae76-900394324ecf","Type":"ContainerStarted","Data":"e49108eba615e1db3e671965a8ef7bb28643912d1c4a6d72e6e3212da5f82ab5"} Dec 01 21:53:18 crc kubenswrapper[4857]: I1201 21:53:18.008131 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3","Type":"ContainerStarted","Data":"f94543767073210665ee75f111efc5037ff585a35ada793039c6626a194e0e0d"} Dec 01 21:53:18 crc kubenswrapper[4857]: I1201 21:53:18.011939 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nhfhn" event={"ID":"32213ad3-4287-4968-9d42-bfeb2ff558d1","Type":"ContainerStarted","Data":"fbaa6096d28d1a177de295a90dddd07fec6fc81ad2a9487c6b5fa9a50b117d8a"} Dec 01 21:53:18 crc kubenswrapper[4857]: I1201 21:53:18.012195 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-nhfhn" Dec 01 21:53:18 crc kubenswrapper[4857]: I1201 21:53:18.043914 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.188407288 podStartE2EDuration="35.043866724s" podCreationTimestamp="2025-12-01 21:52:43 +0000 UTC" firstStartedPulling="2025-12-01 21:52:45.024066933 +0000 UTC m=+1083.514129250" lastFinishedPulling="2025-12-01 21:53:16.879526369 +0000 UTC m=+1115.369588686" observedRunningTime="2025-12-01 21:53:18.037327436 +0000 UTC m=+1116.527389763" watchObservedRunningTime="2025-12-01 21:53:18.043866724 +0000 UTC m=+1116.533929041" Dec 01 21:53:18 crc kubenswrapper[4857]: I1201 21:53:18.064686 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-nhfhn" podStartSLOduration=28.016733338 podStartE2EDuration="32.064669958s" podCreationTimestamp="2025-12-01 21:52:46 +0000 UTC" firstStartedPulling="2025-12-01 21:53:12.86381321 +0000 UTC m=+1111.353875527" lastFinishedPulling="2025-12-01 21:53:16.91174983 +0000 UTC m=+1115.401812147" observedRunningTime="2025-12-01 21:53:18.057604227 +0000 UTC m=+1116.547666544" watchObservedRunningTime="2025-12-01 21:53:18.064669958 +0000 UTC m=+1116.554732275" Dec 01 21:53:19 crc kubenswrapper[4857]: I1201 21:53:19.032520 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gwbmr" event={"ID":"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f","Type":"ContainerStarted","Data":"2e2f5f5b2ed5356ef689ba8b7a29f7ecc2f1ba0637ed7a4407fd557deccfc2f8"} Dec 01 21:53:19 crc kubenswrapper[4857]: I1201 21:53:19.032896 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gwbmr" event={"ID":"5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f","Type":"ContainerStarted","Data":"7f6d94f9f0674d78bef62c1ece9c06f57ddbeda8b32c03a472de327415062988"} Dec 01 21:53:19 crc kubenswrapper[4857]: I1201 21:53:19.062972 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-gwbmr" podStartSLOduration=29.519253941 podStartE2EDuration="33.062957352s" podCreationTimestamp="2025-12-01 21:52:46 +0000 UTC" firstStartedPulling="2025-12-01 21:53:13.365374464 +0000 UTC m=+1111.855436781" lastFinishedPulling="2025-12-01 21:53:16.909077875 +0000 UTC m=+1115.399140192" observedRunningTime="2025-12-01 21:53:19.061995869 +0000 UTC m=+1117.552058196" watchObservedRunningTime="2025-12-01 21:53:19.062957352 +0000 UTC m=+1117.553019669" Dec 01 21:53:20 crc kubenswrapper[4857]: I1201 21:53:20.043272 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:53:20 crc kubenswrapper[4857]: I1201 21:53:20.043787 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:53:21 crc kubenswrapper[4857]: I1201 21:53:21.052823 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b726f947-1deb-4268-ae76-900394324ecf","Type":"ContainerStarted","Data":"a24fa9746f792694ef5317da882b0e29aaf9e667e6cc2d9261513683c8a52a5e"} Dec 01 21:53:21 crc kubenswrapper[4857]: I1201 21:53:21.054235 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"be0a242c-3c02-4d1e-910b-21e74e7c4e43","Type":"ContainerStarted","Data":"73b5306ec343689442e209da468ee0ec0be16e4c60ff6f53124abb5a865e839d"} Dec 01 21:53:21 crc kubenswrapper[4857]: I1201 21:53:21.056425 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3","Type":"ContainerStarted","Data":"5fcac99a8488040cbc3306f414fc8c60987ec2311c067b2616e19387fcf733ad"} Dec 01 21:53:21 crc kubenswrapper[4857]: I1201 21:53:21.058622 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6a290cd2-c321-4cd5-b168-4866c03b135b","Type":"ContainerStarted","Data":"e06bdd4b41e2531261c0be10c0318fe286f401e67bcd1f2b4660cb18a87ddf1d"} Dec 01 21:53:21 crc kubenswrapper[4857]: I1201 21:53:21.059157 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 01 21:53:21 crc kubenswrapper[4857]: I1201 21:53:21.061060 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b3755be2-a089-4259-af43-6e0fbedef24b","Type":"ContainerStarted","Data":"49e0f66360a8aa81f1b578d5360db2f71596ce1a9a631ba7c7dc3a262bc07fc3"} Dec 01 21:53:21 crc kubenswrapper[4857]: I1201 21:53:21.113828 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=24.627387139 podStartE2EDuration="32.113796842s" podCreationTimestamp="2025-12-01 21:52:49 +0000 UTC" firstStartedPulling="2025-12-01 21:53:12.763912891 +0000 UTC m=+1111.253975208" lastFinishedPulling="2025-12-01 21:53:20.250322574 +0000 UTC m=+1118.740384911" observedRunningTime="2025-12-01 21:53:21.090939369 +0000 UTC m=+1119.581001696" watchObservedRunningTime="2025-12-01 21:53:21.113796842 +0000 UTC m=+1119.603859169" Dec 01 21:53:21 crc kubenswrapper[4857]: I1201 21:53:21.153306 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.804528309 podStartE2EDuration="40.153288549s" podCreationTimestamp="2025-12-01 21:52:41 +0000 UTC" firstStartedPulling="2025-12-01 21:52:42.997213079 +0000 UTC m=+1081.487275396" lastFinishedPulling="2025-12-01 21:53:20.345973319 +0000 UTC m=+1118.836035636" observedRunningTime="2025-12-01 21:53:21.152646833 +0000 UTC m=+1119.642709160" watchObservedRunningTime="2025-12-01 21:53:21.153288549 +0000 UTC m=+1119.643350866" Dec 01 21:53:21 crc kubenswrapper[4857]: I1201 21:53:21.177652 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 01 21:53:21 crc kubenswrapper[4857]: I1201 21:53:21.177723 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 01 21:53:21 crc kubenswrapper[4857]: I1201 21:53:21.230696 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 01 21:53:21 crc kubenswrapper[4857]: I1201 21:53:21.292671 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=26.417720147 podStartE2EDuration="33.292653623s" podCreationTimestamp="2025-12-01 21:52:48 +0000 UTC" firstStartedPulling="2025-12-01 21:53:13.466876692 +0000 UTC m=+1111.956939009" lastFinishedPulling="2025-12-01 21:53:20.341810158 +0000 UTC m=+1118.831872485" observedRunningTime="2025-12-01 21:53:21.177456104 +0000 UTC m=+1119.667518421" watchObservedRunningTime="2025-12-01 21:53:21.292653623 +0000 UTC m=+1119.782715940" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.135094 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.402419 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mw6gs"] Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.455350 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-lqjrc"] Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.456589 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.458650 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.461984 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f5g59"] Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.463366 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.466034 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.474812 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lqjrc"] Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.481318 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f5g59"] Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.556301 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-config\") pod \"dnsmasq-dns-7fd796d7df-f5g59\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.556339 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7aab234d-a569-48bb-b75b-6a4ecf993df7-ovn-rundir\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.556357 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7aab234d-a569-48bb-b75b-6a4ecf993df7-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.556380 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7aab234d-a569-48bb-b75b-6a4ecf993df7-ovs-rundir\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.556400 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aab234d-a569-48bb-b75b-6a4ecf993df7-config\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.556422 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aab234d-a569-48bb-b75b-6a4ecf993df7-combined-ca-bundle\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.556448 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-f5g59\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.556491 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-f5g59\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.556510 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz9nt\" (UniqueName: \"kubernetes.io/projected/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-kube-api-access-kz9nt\") pod \"dnsmasq-dns-7fd796d7df-f5g59\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.556577 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x6zz\" (UniqueName: \"kubernetes.io/projected/7aab234d-a569-48bb-b75b-6a4ecf993df7-kube-api-access-8x6zz\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.590803 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.649157 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.658422 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-f5g59\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.658469 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz9nt\" (UniqueName: \"kubernetes.io/projected/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-kube-api-access-kz9nt\") pod \"dnsmasq-dns-7fd796d7df-f5g59\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.658532 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x6zz\" (UniqueName: \"kubernetes.io/projected/7aab234d-a569-48bb-b75b-6a4ecf993df7-kube-api-access-8x6zz\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.658565 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-config\") pod \"dnsmasq-dns-7fd796d7df-f5g59\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.658588 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7aab234d-a569-48bb-b75b-6a4ecf993df7-ovn-rundir\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.658607 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7aab234d-a569-48bb-b75b-6a4ecf993df7-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.658625 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7aab234d-a569-48bb-b75b-6a4ecf993df7-ovs-rundir\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.658644 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aab234d-a569-48bb-b75b-6a4ecf993df7-config\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.658664 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aab234d-a569-48bb-b75b-6a4ecf993df7-combined-ca-bundle\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.658683 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-f5g59\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.659585 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-f5g59\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.661132 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-f5g59\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.661964 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-config\") pod \"dnsmasq-dns-7fd796d7df-f5g59\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.662225 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7aab234d-a569-48bb-b75b-6a4ecf993df7-ovn-rundir\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.666581 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aab234d-a569-48bb-b75b-6a4ecf993df7-config\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.666725 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7aab234d-a569-48bb-b75b-6a4ecf993df7-ovs-rundir\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.668840 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7aab234d-a569-48bb-b75b-6a4ecf993df7-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.671307 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aab234d-a569-48bb-b75b-6a4ecf993df7-combined-ca-bundle\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.683310 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x6zz\" (UniqueName: \"kubernetes.io/projected/7aab234d-a569-48bb-b75b-6a4ecf993df7-kube-api-access-8x6zz\") pod \"ovn-controller-metrics-lqjrc\" (UID: \"7aab234d-a569-48bb-b75b-6a4ecf993df7\") " pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.695950 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz9nt\" (UniqueName: \"kubernetes.io/projected/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-kube-api-access-kz9nt\") pod \"dnsmasq-dns-7fd796d7df-f5g59\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.794904 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-l7k8n"] Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.846876 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-g6z4j"] Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.858033 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.862574 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.872522 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lqjrc" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.873327 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-g6z4j"] Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.874690 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.935057 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.972406 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.972664 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-config\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.972871 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2smx7\" (UniqueName: \"kubernetes.io/projected/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-kube-api-access-2smx7\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.972960 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:22 crc kubenswrapper[4857]: I1201 21:53:22.973068 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.074955 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0253178-b993-476c-8d26-844ed2851e06-config\") pod \"b0253178-b993-476c-8d26-844ed2851e06\" (UID: \"b0253178-b993-476c-8d26-844ed2851e06\") " Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.081426 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0253178-b993-476c-8d26-844ed2851e06-config" (OuterVolumeSpecName: "config") pod "b0253178-b993-476c-8d26-844ed2851e06" (UID: "b0253178-b993-476c-8d26-844ed2851e06"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.083375 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fmsw\" (UniqueName: \"kubernetes.io/projected/b0253178-b993-476c-8d26-844ed2851e06-kube-api-access-6fmsw\") pod \"b0253178-b993-476c-8d26-844ed2851e06\" (UID: \"b0253178-b993-476c-8d26-844ed2851e06\") " Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.083516 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0253178-b993-476c-8d26-844ed2851e06-dns-svc\") pod \"b0253178-b993-476c-8d26-844ed2851e06\" (UID: \"b0253178-b993-476c-8d26-844ed2851e06\") " Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.085622 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-config\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.085825 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2smx7\" (UniqueName: \"kubernetes.io/projected/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-kube-api-access-2smx7\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.085864 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.085919 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.086001 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.086175 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0253178-b993-476c-8d26-844ed2851e06-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.087025 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-config\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.087319 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.087399 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0253178-b993-476c-8d26-844ed2851e06-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b0253178-b993-476c-8d26-844ed2851e06" (UID: "b0253178-b993-476c-8d26-844ed2851e06"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.087574 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.088568 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.090279 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0253178-b993-476c-8d26-844ed2851e06-kube-api-access-6fmsw" (OuterVolumeSpecName: "kube-api-access-6fmsw") pod "b0253178-b993-476c-8d26-844ed2851e06" (UID: "b0253178-b993-476c-8d26-844ed2851e06"). InnerVolumeSpecName "kube-api-access-6fmsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.090946 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" event={"ID":"b0253178-b993-476c-8d26-844ed2851e06","Type":"ContainerDied","Data":"2fe1fe74fc90c941a3eabe087b2f2bb2b88fa7719d9a963b78d289ebb4d69302"} Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.091095 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-mw6gs" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.091923 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.119200 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2smx7\" (UniqueName: \"kubernetes.io/projected/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-kube-api-access-2smx7\") pod \"dnsmasq-dns-86db49b7ff-g6z4j\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.169293 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.171108 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.180098 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mw6gs"] Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.187513 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0253178-b993-476c-8d26-844ed2851e06-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.187544 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fmsw\" (UniqueName: \"kubernetes.io/projected/b0253178-b993-476c-8d26-844ed2851e06-kube-api-access-6fmsw\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.209882 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mw6gs"] Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.248901 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.289377 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69b986d7-590f-4fa1-b2ab-2bef95fe745e-config\") pod \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\" (UID: \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\") " Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.289418 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69b986d7-590f-4fa1-b2ab-2bef95fe745e-dns-svc\") pod \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\" (UID: \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\") " Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.289522 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxtst\" (UniqueName: \"kubernetes.io/projected/69b986d7-590f-4fa1-b2ab-2bef95fe745e-kube-api-access-fxtst\") pod \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\" (UID: \"69b986d7-590f-4fa1-b2ab-2bef95fe745e\") " Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.292518 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b986d7-590f-4fa1-b2ab-2bef95fe745e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "69b986d7-590f-4fa1-b2ab-2bef95fe745e" (UID: "69b986d7-590f-4fa1-b2ab-2bef95fe745e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.293662 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b986d7-590f-4fa1-b2ab-2bef95fe745e-config" (OuterVolumeSpecName: "config") pod "69b986d7-590f-4fa1-b2ab-2bef95fe745e" (UID: "69b986d7-590f-4fa1-b2ab-2bef95fe745e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.298647 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69b986d7-590f-4fa1-b2ab-2bef95fe745e-kube-api-access-fxtst" (OuterVolumeSpecName: "kube-api-access-fxtst") pod "69b986d7-590f-4fa1-b2ab-2bef95fe745e" (UID: "69b986d7-590f-4fa1-b2ab-2bef95fe745e"). InnerVolumeSpecName "kube-api-access-fxtst". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.392417 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69b986d7-590f-4fa1-b2ab-2bef95fe745e-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.392455 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69b986d7-590f-4fa1-b2ab-2bef95fe745e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.392467 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxtst\" (UniqueName: \"kubernetes.io/projected/69b986d7-590f-4fa1-b2ab-2bef95fe745e-kube-api-access-fxtst\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.413559 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f5g59"] Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.581916 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lqjrc"] Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.655071 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.656641 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.666293 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.666775 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.666945 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-hjmxv" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.667154 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.673921 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.803422 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/467e700f-9a49-4c3d-8108-9bb78cc31495-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.803472 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/467e700f-9a49-4c3d-8108-9bb78cc31495-config\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.803556 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lwd6\" (UniqueName: \"kubernetes.io/projected/467e700f-9a49-4c3d-8108-9bb78cc31495-kube-api-access-8lwd6\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.803594 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/467e700f-9a49-4c3d-8108-9bb78cc31495-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.803629 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/467e700f-9a49-4c3d-8108-9bb78cc31495-scripts\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.803645 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/467e700f-9a49-4c3d-8108-9bb78cc31495-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.803665 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467e700f-9a49-4c3d-8108-9bb78cc31495-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.846617 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0253178-b993-476c-8d26-844ed2851e06" path="/var/lib/kubelet/pods/b0253178-b993-476c-8d26-844ed2851e06/volumes" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.904937 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lwd6\" (UniqueName: \"kubernetes.io/projected/467e700f-9a49-4c3d-8108-9bb78cc31495-kube-api-access-8lwd6\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.904993 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/467e700f-9a49-4c3d-8108-9bb78cc31495-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.905034 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/467e700f-9a49-4c3d-8108-9bb78cc31495-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.905072 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/467e700f-9a49-4c3d-8108-9bb78cc31495-scripts\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.905098 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467e700f-9a49-4c3d-8108-9bb78cc31495-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.905208 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/467e700f-9a49-4c3d-8108-9bb78cc31495-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.905240 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/467e700f-9a49-4c3d-8108-9bb78cc31495-config\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.906203 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/467e700f-9a49-4c3d-8108-9bb78cc31495-scripts\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.907182 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/467e700f-9a49-4c3d-8108-9bb78cc31495-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.907650 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/467e700f-9a49-4c3d-8108-9bb78cc31495-config\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.919497 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467e700f-9a49-4c3d-8108-9bb78cc31495-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.919518 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/467e700f-9a49-4c3d-8108-9bb78cc31495-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.919497 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/467e700f-9a49-4c3d-8108-9bb78cc31495-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.922509 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-g6z4j"] Dec 01 21:53:23 crc kubenswrapper[4857]: I1201 21:53:23.926315 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lwd6\" (UniqueName: \"kubernetes.io/projected/467e700f-9a49-4c3d-8108-9bb78cc31495-kube-api-access-8lwd6\") pod \"ovn-northd-0\" (UID: \"467e700f-9a49-4c3d-8108-9bb78cc31495\") " pod="openstack/ovn-northd-0" Dec 01 21:53:23 crc kubenswrapper[4857]: W1201 21:53:23.926428 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e724c46_b6d3_4e20_910b_8a7e417fd0cb.slice/crio-4845bf4381cfcdcf9db89c765f6fca6219a61c2db671b8f3abb1b9e1551d49a6 WatchSource:0}: Error finding container 4845bf4381cfcdcf9db89c765f6fca6219a61c2db671b8f3abb1b9e1551d49a6: Status 404 returned error can't find the container with id 4845bf4381cfcdcf9db89c765f6fca6219a61c2db671b8f3abb1b9e1551d49a6 Dec 01 21:53:24 crc kubenswrapper[4857]: I1201 21:53:24.024875 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 21:53:24 crc kubenswrapper[4857]: I1201 21:53:24.096079 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 21:53:24 crc kubenswrapper[4857]: I1201 21:53:24.101847 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lqjrc" event={"ID":"7aab234d-a569-48bb-b75b-6a4ecf993df7","Type":"ContainerStarted","Data":"3c52311115a4f53521254ea4bfdc8b7330595dac201162568121cd7205b83904"} Dec 01 21:53:24 crc kubenswrapper[4857]: I1201 21:53:24.102797 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" event={"ID":"69b986d7-590f-4fa1-b2ab-2bef95fe745e","Type":"ContainerDied","Data":"e421e0aff90b3152383b2d4260fcd344e57348be1a01b7b369c892f9f368a478"} Dec 01 21:53:24 crc kubenswrapper[4857]: I1201 21:53:24.102900 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-l7k8n" Dec 01 21:53:24 crc kubenswrapper[4857]: I1201 21:53:24.105285 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" event={"ID":"3cc1f7a9-0f68-486f-8d77-3cec7f184f17","Type":"ContainerStarted","Data":"d796066d649336096db24e625c89208d82d270a2d33709a200c9e0d0047a14a4"} Dec 01 21:53:24 crc kubenswrapper[4857]: I1201 21:53:24.107822 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" event={"ID":"8e724c46-b6d3-4e20-910b-8a7e417fd0cb","Type":"ContainerStarted","Data":"4845bf4381cfcdcf9db89c765f6fca6219a61c2db671b8f3abb1b9e1551d49a6"} Dec 01 21:53:24 crc kubenswrapper[4857]: I1201 21:53:24.171617 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-l7k8n"] Dec 01 21:53:24 crc kubenswrapper[4857]: I1201 21:53:24.180089 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-l7k8n"] Dec 01 21:53:24 crc kubenswrapper[4857]: I1201 21:53:24.621236 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 21:53:24 crc kubenswrapper[4857]: W1201 21:53:24.636919 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod467e700f_9a49_4c3d_8108_9bb78cc31495.slice/crio-cb4f572aa3de4e9ba70dcba531c856210d42698d18f207328c093380912ea4ab WatchSource:0}: Error finding container cb4f572aa3de4e9ba70dcba531c856210d42698d18f207328c093380912ea4ab: Status 404 returned error can't find the container with id cb4f572aa3de4e9ba70dcba531c856210d42698d18f207328c093380912ea4ab Dec 01 21:53:25 crc kubenswrapper[4857]: I1201 21:53:25.113467 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" event={"ID":"3cc1f7a9-0f68-486f-8d77-3cec7f184f17","Type":"ContainerStarted","Data":"0dc71dd3d5203b41dc2d657ac7767e47e6d4be47c066d105bad70463574a81b9"} Dec 01 21:53:25 crc kubenswrapper[4857]: I1201 21:53:25.114713 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"467e700f-9a49-4c3d-8108-9bb78cc31495","Type":"ContainerStarted","Data":"cb4f572aa3de4e9ba70dcba531c856210d42698d18f207328c093380912ea4ab"} Dec 01 21:53:25 crc kubenswrapper[4857]: I1201 21:53:25.116219 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lqjrc" event={"ID":"7aab234d-a569-48bb-b75b-6a4ecf993df7","Type":"ContainerStarted","Data":"2b40de7d4b361418c4f5ba613814de457681bd9865e9fb0863075aeec4050274"} Dec 01 21:53:25 crc kubenswrapper[4857]: I1201 21:53:25.180626 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-lqjrc" podStartSLOduration=3.18060657 podStartE2EDuration="3.18060657s" podCreationTimestamp="2025-12-01 21:53:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:53:25.176780367 +0000 UTC m=+1123.666842684" watchObservedRunningTime="2025-12-01 21:53:25.18060657 +0000 UTC m=+1123.670668887" Dec 01 21:53:25 crc kubenswrapper[4857]: I1201 21:53:25.847027 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69b986d7-590f-4fa1-b2ab-2bef95fe745e" path="/var/lib/kubelet/pods/69b986d7-590f-4fa1-b2ab-2bef95fe745e/volumes" Dec 01 21:53:26 crc kubenswrapper[4857]: I1201 21:53:26.128763 4857 generic.go:334] "Generic (PLEG): container finished" podID="3cc1f7a9-0f68-486f-8d77-3cec7f184f17" containerID="0dc71dd3d5203b41dc2d657ac7767e47e6d4be47c066d105bad70463574a81b9" exitCode=0 Dec 01 21:53:26 crc kubenswrapper[4857]: I1201 21:53:26.128870 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" event={"ID":"3cc1f7a9-0f68-486f-8d77-3cec7f184f17","Type":"ContainerDied","Data":"0dc71dd3d5203b41dc2d657ac7767e47e6d4be47c066d105bad70463574a81b9"} Dec 01 21:53:26 crc kubenswrapper[4857]: I1201 21:53:26.132639 4857 generic.go:334] "Generic (PLEG): container finished" podID="8e724c46-b6d3-4e20-910b-8a7e417fd0cb" containerID="fa02bc89e5d41b2cf8e97787bed6fedf04aafe584d76281cdf7495bdf95ef3fc" exitCode=0 Dec 01 21:53:26 crc kubenswrapper[4857]: I1201 21:53:26.132934 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" event={"ID":"8e724c46-b6d3-4e20-910b-8a7e417fd0cb","Type":"ContainerDied","Data":"fa02bc89e5d41b2cf8e97787bed6fedf04aafe584d76281cdf7495bdf95ef3fc"} Dec 01 21:53:26 crc kubenswrapper[4857]: I1201 21:53:26.137725 4857 generic.go:334] "Generic (PLEG): container finished" podID="be0a242c-3c02-4d1e-910b-21e74e7c4e43" containerID="73b5306ec343689442e209da468ee0ec0be16e4c60ff6f53124abb5a865e839d" exitCode=0 Dec 01 21:53:26 crc kubenswrapper[4857]: I1201 21:53:26.137772 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"be0a242c-3c02-4d1e-910b-21e74e7c4e43","Type":"ContainerDied","Data":"73b5306ec343689442e209da468ee0ec0be16e4c60ff6f53124abb5a865e839d"} Dec 01 21:53:26 crc kubenswrapper[4857]: I1201 21:53:26.711265 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.147481 4857 generic.go:334] "Generic (PLEG): container finished" podID="b3755be2-a089-4259-af43-6e0fbedef24b" containerID="49e0f66360a8aa81f1b578d5360db2f71596ce1a9a631ba7c7dc3a262bc07fc3" exitCode=0 Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.147577 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b3755be2-a089-4259-af43-6e0fbedef24b","Type":"ContainerDied","Data":"49e0f66360a8aa81f1b578d5360db2f71596ce1a9a631ba7c7dc3a262bc07fc3"} Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.149716 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"467e700f-9a49-4c3d-8108-9bb78cc31495","Type":"ContainerStarted","Data":"9ba824df6083586aecf2d54e08216e872602c3f4ad211a5a4554b24026d15614"} Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.149754 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"467e700f-9a49-4c3d-8108-9bb78cc31495","Type":"ContainerStarted","Data":"0f96f87a489fc21ba9416cfd19e1f9cf67b43af44afa4c7079f377b25d0632a7"} Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.150309 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.153265 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"be0a242c-3c02-4d1e-910b-21e74e7c4e43","Type":"ContainerStarted","Data":"4b2a24a262ce528cc58536b12cac89bb68b03fba8904f242c03241d22bbcf1e0"} Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.157234 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" event={"ID":"3cc1f7a9-0f68-486f-8d77-3cec7f184f17","Type":"ContainerStarted","Data":"2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df"} Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.157814 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.160895 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" event={"ID":"8e724c46-b6d3-4e20-910b-8a7e417fd0cb","Type":"ContainerStarted","Data":"d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445"} Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.161428 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.200927 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" podStartSLOduration=4.014654037 podStartE2EDuration="5.200489692s" podCreationTimestamp="2025-12-01 21:53:22 +0000 UTC" firstStartedPulling="2025-12-01 21:53:23.937424126 +0000 UTC m=+1122.427486443" lastFinishedPulling="2025-12-01 21:53:25.123259781 +0000 UTC m=+1123.613322098" observedRunningTime="2025-12-01 21:53:27.194766934 +0000 UTC m=+1125.684829261" watchObservedRunningTime="2025-12-01 21:53:27.200489692 +0000 UTC m=+1125.690552009" Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.224302 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=10.742596247 podStartE2EDuration="48.224280578s" podCreationTimestamp="2025-12-01 21:52:39 +0000 UTC" firstStartedPulling="2025-12-01 21:52:42.759337448 +0000 UTC m=+1081.249399765" lastFinishedPulling="2025-12-01 21:53:20.241021749 +0000 UTC m=+1118.731084096" observedRunningTime="2025-12-01 21:53:27.217801331 +0000 UTC m=+1125.707863648" watchObservedRunningTime="2025-12-01 21:53:27.224280578 +0000 UTC m=+1125.714342895" Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.264163 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" podStartSLOduration=3.878541001 podStartE2EDuration="5.264141163s" podCreationTimestamp="2025-12-01 21:53:22 +0000 UTC" firstStartedPulling="2025-12-01 21:53:23.488491526 +0000 UTC m=+1121.978553843" lastFinishedPulling="2025-12-01 21:53:24.874091688 +0000 UTC m=+1123.364154005" observedRunningTime="2025-12-01 21:53:27.239308962 +0000 UTC m=+1125.729371289" watchObservedRunningTime="2025-12-01 21:53:27.264141163 +0000 UTC m=+1125.754203480" Dec 01 21:53:27 crc kubenswrapper[4857]: I1201 21:53:27.267713 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.077551299 podStartE2EDuration="4.267693449s" podCreationTimestamp="2025-12-01 21:53:23 +0000 UTC" firstStartedPulling="2025-12-01 21:53:24.641123676 +0000 UTC m=+1123.131185983" lastFinishedPulling="2025-12-01 21:53:25.831265816 +0000 UTC m=+1124.321328133" observedRunningTime="2025-12-01 21:53:27.259539302 +0000 UTC m=+1125.749601639" watchObservedRunningTime="2025-12-01 21:53:27.267693449 +0000 UTC m=+1125.757755766" Dec 01 21:53:28 crc kubenswrapper[4857]: I1201 21:53:28.173725 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b3755be2-a089-4259-af43-6e0fbedef24b","Type":"ContainerStarted","Data":"bcd8bb421fedfbbec458815d4cfa9463c526976ad0d3c6ad38412394ffedec97"} Dec 01 21:53:28 crc kubenswrapper[4857]: I1201 21:53:28.214242 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371987.640568 podStartE2EDuration="49.214207249s" podCreationTimestamp="2025-12-01 21:52:39 +0000 UTC" firstStartedPulling="2025-12-01 21:52:40.932327618 +0000 UTC m=+1079.422389935" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:53:28.204341 +0000 UTC m=+1126.694403317" watchObservedRunningTime="2025-12-01 21:53:28.214207249 +0000 UTC m=+1126.704269566" Dec 01 21:53:30 crc kubenswrapper[4857]: I1201 21:53:30.526190 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 01 21:53:30 crc kubenswrapper[4857]: I1201 21:53:30.526610 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 01 21:53:31 crc kubenswrapper[4857]: I1201 21:53:31.674623 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 01 21:53:31 crc kubenswrapper[4857]: I1201 21:53:31.674682 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 01 21:53:32 crc kubenswrapper[4857]: I1201 21:53:32.877023 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:33 crc kubenswrapper[4857]: I1201 21:53:33.251248 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:33 crc kubenswrapper[4857]: I1201 21:53:33.308937 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f5g59"] Dec 01 21:53:33 crc kubenswrapper[4857]: I1201 21:53:33.309159 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" podUID="3cc1f7a9-0f68-486f-8d77-3cec7f184f17" containerName="dnsmasq-dns" containerID="cri-o://2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df" gracePeriod=10 Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.235792 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-tcxjl"] Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.237574 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.249837 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tcxjl"] Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.403805 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bnh8\" (UniqueName: \"kubernetes.io/projected/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-kube-api-access-6bnh8\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.404218 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-dns-svc\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.404294 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.404363 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.404508 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-config\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.506338 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-dns-svc\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.506428 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.506482 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.506593 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-config\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.506673 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bnh8\" (UniqueName: \"kubernetes.io/projected/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-kube-api-access-6bnh8\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.507300 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-dns-svc\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.507506 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.507988 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-config\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.507994 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.528931 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bnh8\" (UniqueName: \"kubernetes.io/projected/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-kube-api-access-6bnh8\") pod \"dnsmasq-dns-698758b865-tcxjl\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:34 crc kubenswrapper[4857]: I1201 21:53:34.566202 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.030055 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tcxjl"] Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.145813 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.228627 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-dns-svc\") pod \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.231250 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kz9nt\" (UniqueName: \"kubernetes.io/projected/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-kube-api-access-kz9nt\") pod \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.276981 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-config\") pod \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.277167 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-ovsdbserver-nb\") pod \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\" (UID: \"3cc1f7a9-0f68-486f-8d77-3cec7f184f17\") " Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.283989 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-kube-api-access-kz9nt" (OuterVolumeSpecName: "kube-api-access-kz9nt") pod "3cc1f7a9-0f68-486f-8d77-3cec7f184f17" (UID: "3cc1f7a9-0f68-486f-8d77-3cec7f184f17"). InnerVolumeSpecName "kube-api-access-kz9nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.288113 4857 generic.go:334] "Generic (PLEG): container finished" podID="3cc1f7a9-0f68-486f-8d77-3cec7f184f17" containerID="2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df" exitCode=0 Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.288404 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.288452 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" event={"ID":"3cc1f7a9-0f68-486f-8d77-3cec7f184f17","Type":"ContainerDied","Data":"2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df"} Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.288537 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-f5g59" event={"ID":"3cc1f7a9-0f68-486f-8d77-3cec7f184f17","Type":"ContainerDied","Data":"d796066d649336096db24e625c89208d82d270a2d33709a200c9e0d0047a14a4"} Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.288563 4857 scope.go:117] "RemoveContainer" containerID="2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.289511 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tcxjl" event={"ID":"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0","Type":"ContainerStarted","Data":"647237fa82377c40370d08596511110ba06e34acdca35bf7ed18da72773b2cc5"} Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.309298 4857 scope.go:117] "RemoveContainer" containerID="0dc71dd3d5203b41dc2d657ac7767e47e6d4be47c066d105bad70463574a81b9" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.324617 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-config" (OuterVolumeSpecName: "config") pod "3cc1f7a9-0f68-486f-8d77-3cec7f184f17" (UID: "3cc1f7a9-0f68-486f-8d77-3cec7f184f17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.324720 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3cc1f7a9-0f68-486f-8d77-3cec7f184f17" (UID: "3cc1f7a9-0f68-486f-8d77-3cec7f184f17"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.335104 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3cc1f7a9-0f68-486f-8d77-3cec7f184f17" (UID: "3cc1f7a9-0f68-486f-8d77-3cec7f184f17"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.339994 4857 scope.go:117] "RemoveContainer" containerID="2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df" Dec 01 21:53:35 crc kubenswrapper[4857]: E1201 21:53:35.340520 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df\": container with ID starting with 2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df not found: ID does not exist" containerID="2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.340563 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df"} err="failed to get container status \"2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df\": rpc error: code = NotFound desc = could not find container \"2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df\": container with ID starting with 2de501cb4a4ba2d7a46fe06eb02082b484a0a074d7839323345901559e0e82df not found: ID does not exist" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.340598 4857 scope.go:117] "RemoveContainer" containerID="0dc71dd3d5203b41dc2d657ac7767e47e6d4be47c066d105bad70463574a81b9" Dec 01 21:53:35 crc kubenswrapper[4857]: E1201 21:53:35.341118 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dc71dd3d5203b41dc2d657ac7767e47e6d4be47c066d105bad70463574a81b9\": container with ID starting with 0dc71dd3d5203b41dc2d657ac7767e47e6d4be47c066d105bad70463574a81b9 not found: ID does not exist" containerID="0dc71dd3d5203b41dc2d657ac7767e47e6d4be47c066d105bad70463574a81b9" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.341145 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dc71dd3d5203b41dc2d657ac7767e47e6d4be47c066d105bad70463574a81b9"} err="failed to get container status \"0dc71dd3d5203b41dc2d657ac7767e47e6d4be47c066d105bad70463574a81b9\": rpc error: code = NotFound desc = could not find container \"0dc71dd3d5203b41dc2d657ac7767e47e6d4be47c066d105bad70463574a81b9\": container with ID starting with 0dc71dd3d5203b41dc2d657ac7767e47e6d4be47c066d105bad70463574a81b9 not found: ID does not exist" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.354596 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 01 21:53:35 crc kubenswrapper[4857]: E1201 21:53:35.354896 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cc1f7a9-0f68-486f-8d77-3cec7f184f17" containerName="dnsmasq-dns" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.354912 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cc1f7a9-0f68-486f-8d77-3cec7f184f17" containerName="dnsmasq-dns" Dec 01 21:53:35 crc kubenswrapper[4857]: E1201 21:53:35.354935 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cc1f7a9-0f68-486f-8d77-3cec7f184f17" containerName="init" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.354943 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cc1f7a9-0f68-486f-8d77-3cec7f184f17" containerName="init" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.355223 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cc1f7a9-0f68-486f-8d77-3cec7f184f17" containerName="dnsmasq-dns" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.360790 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.362698 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.362806 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.363430 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-6vlfd" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.374474 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.378896 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.378912 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.378923 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.378933 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kz9nt\" (UniqueName: \"kubernetes.io/projected/3cc1f7a9-0f68-486f-8d77-3cec7f184f17-kube-api-access-kz9nt\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.389778 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.480480 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3ea80def-31e0-4cc2-9f15-26ee35c62641-cache\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.480570 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.480670 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3ea80def-31e0-4cc2-9f15-26ee35c62641-lock\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.480756 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.480826 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlz2n\" (UniqueName: \"kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-kube-api-access-rlz2n\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.582416 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlz2n\" (UniqueName: \"kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-kube-api-access-rlz2n\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.582733 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3ea80def-31e0-4cc2-9f15-26ee35c62641-cache\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.582907 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.583021 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3ea80def-31e0-4cc2-9f15-26ee35c62641-lock\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.583139 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.583354 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3ea80def-31e0-4cc2-9f15-26ee35c62641-cache\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: E1201 21:53:35.583369 4857 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 21:53:35 crc kubenswrapper[4857]: E1201 21:53:35.583427 4857 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 21:53:35 crc kubenswrapper[4857]: E1201 21:53:35.583466 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift podName:3ea80def-31e0-4cc2-9f15-26ee35c62641 nodeName:}" failed. No retries permitted until 2025-12-01 21:53:36.083452205 +0000 UTC m=+1134.573514522 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift") pod "swift-storage-0" (UID: "3ea80def-31e0-4cc2-9f15-26ee35c62641") : configmap "swift-ring-files" not found Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.583712 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3ea80def-31e0-4cc2-9f15-26ee35c62641-lock\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.584228 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.601992 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlz2n\" (UniqueName: \"kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-kube-api-access-rlz2n\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.618376 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.676334 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f5g59"] Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.681970 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f5g59"] Dec 01 21:53:35 crc kubenswrapper[4857]: I1201 21:53:35.845272 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cc1f7a9-0f68-486f-8d77-3cec7f184f17" path="/var/lib/kubelet/pods/3cc1f7a9-0f68-486f-8d77-3cec7f184f17/volumes" Dec 01 21:53:36 crc kubenswrapper[4857]: I1201 21:53:36.091429 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:36 crc kubenswrapper[4857]: E1201 21:53:36.091639 4857 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 21:53:36 crc kubenswrapper[4857]: E1201 21:53:36.091969 4857 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 21:53:36 crc kubenswrapper[4857]: E1201 21:53:36.092116 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift podName:3ea80def-31e0-4cc2-9f15-26ee35c62641 nodeName:}" failed. No retries permitted until 2025-12-01 21:53:37.092093792 +0000 UTC m=+1135.582156109 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift") pod "swift-storage-0" (UID: "3ea80def-31e0-4cc2-9f15-26ee35c62641") : configmap "swift-ring-files" not found Dec 01 21:53:36 crc kubenswrapper[4857]: I1201 21:53:36.301896 4857 generic.go:334] "Generic (PLEG): container finished" podID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerID="eadafa0db16cde34081bb8dd379cae4f4e309967f44216a5d4139184f11a3042" exitCode=0 Dec 01 21:53:36 crc kubenswrapper[4857]: I1201 21:53:36.302112 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tcxjl" event={"ID":"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0","Type":"ContainerDied","Data":"eadafa0db16cde34081bb8dd379cae4f4e309967f44216a5d4139184f11a3042"} Dec 01 21:53:36 crc kubenswrapper[4857]: I1201 21:53:36.862057 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 01 21:53:36 crc kubenswrapper[4857]: I1201 21:53:36.966532 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 01 21:53:37 crc kubenswrapper[4857]: I1201 21:53:37.110258 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:37 crc kubenswrapper[4857]: E1201 21:53:37.110497 4857 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 21:53:37 crc kubenswrapper[4857]: E1201 21:53:37.110645 4857 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 21:53:37 crc kubenswrapper[4857]: E1201 21:53:37.110702 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift podName:3ea80def-31e0-4cc2-9f15-26ee35c62641 nodeName:}" failed. No retries permitted until 2025-12-01 21:53:39.110686727 +0000 UTC m=+1137.600749044 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift") pod "swift-storage-0" (UID: "3ea80def-31e0-4cc2-9f15-26ee35c62641") : configmap "swift-ring-files" not found Dec 01 21:53:37 crc kubenswrapper[4857]: I1201 21:53:37.312652 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tcxjl" event={"ID":"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0","Type":"ContainerStarted","Data":"243f268abfc250a13a95a74cf5150ef217a473f49764d58f80409ac94ca5c64c"} Dec 01 21:53:37 crc kubenswrapper[4857]: I1201 21:53:37.349165 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-tcxjl" podStartSLOduration=3.349138471 podStartE2EDuration="3.349138471s" podCreationTimestamp="2025-12-01 21:53:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:53:37.331421122 +0000 UTC m=+1135.821483449" watchObservedRunningTime="2025-12-01 21:53:37.349138471 +0000 UTC m=+1135.839200788" Dec 01 21:53:38 crc kubenswrapper[4857]: I1201 21:53:38.320505 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:38 crc kubenswrapper[4857]: I1201 21:53:38.732825 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 01 21:53:38 crc kubenswrapper[4857]: I1201 21:53:38.812548 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.147669 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:39 crc kubenswrapper[4857]: E1201 21:53:39.147933 4857 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 21:53:39 crc kubenswrapper[4857]: E1201 21:53:39.147986 4857 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 21:53:39 crc kubenswrapper[4857]: E1201 21:53:39.148078 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift podName:3ea80def-31e0-4cc2-9f15-26ee35c62641 nodeName:}" failed. No retries permitted until 2025-12-01 21:53:43.148056152 +0000 UTC m=+1141.638118469 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift") pod "swift-storage-0" (UID: "3ea80def-31e0-4cc2-9f15-26ee35c62641") : configmap "swift-ring-files" not found Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.158664 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.519471 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-xptvd"] Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.520746 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.523342 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.523636 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.523820 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.567568 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-q5szn"] Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.569781 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.579778 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-xptvd"] Dec 01 21:53:39 crc kubenswrapper[4857]: E1201 21:53:39.580729 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-hv4tz ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-xptvd" podUID="db2d0de5-6765-4080-af32-7e98182627a1" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.605099 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-q5szn"] Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.638162 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-xptvd"] Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.659868 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-swiftconf\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.659928 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-combined-ca-bundle\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.659952 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-dispersionconf\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.660000 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db2d0de5-6765-4080-af32-7e98182627a1-ring-data-devices\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.660054 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db2d0de5-6765-4080-af32-7e98182627a1-scripts\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.660173 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv4tz\" (UniqueName: \"kubernetes.io/projected/db2d0de5-6765-4080-af32-7e98182627a1-kube-api-access-hv4tz\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.660205 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db2d0de5-6765-4080-af32-7e98182627a1-etc-swift\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.761713 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db2d0de5-6765-4080-af32-7e98182627a1-ring-data-devices\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.761813 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-ring-data-devices\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.761861 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db2d0de5-6765-4080-af32-7e98182627a1-scripts\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.761946 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-dispersionconf\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.761997 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv4tz\" (UniqueName: \"kubernetes.io/projected/db2d0de5-6765-4080-af32-7e98182627a1-kube-api-access-hv4tz\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.762075 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-combined-ca-bundle\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.762128 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db2d0de5-6765-4080-af32-7e98182627a1-etc-swift\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.762186 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-swiftconf\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.762241 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-swiftconf\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.762328 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhcln\" (UniqueName: \"kubernetes.io/projected/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-kube-api-access-rhcln\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.762379 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-combined-ca-bundle\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.762497 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-dispersionconf\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.762551 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-etc-swift\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.762571 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-scripts\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.763374 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db2d0de5-6765-4080-af32-7e98182627a1-etc-swift\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.763406 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db2d0de5-6765-4080-af32-7e98182627a1-scripts\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.763380 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db2d0de5-6765-4080-af32-7e98182627a1-ring-data-devices\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.772492 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-combined-ca-bundle\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.773553 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-dispersionconf\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.777552 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-swiftconf\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.781513 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv4tz\" (UniqueName: \"kubernetes.io/projected/db2d0de5-6765-4080-af32-7e98182627a1-kube-api-access-hv4tz\") pod \"swift-ring-rebalance-xptvd\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.864284 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhcln\" (UniqueName: \"kubernetes.io/projected/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-kube-api-access-rhcln\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.864337 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-etc-swift\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.864353 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-scripts\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.864422 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-ring-data-devices\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.864487 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-dispersionconf\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.864524 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-combined-ca-bundle\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.864545 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-swiftconf\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.864948 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-etc-swift\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.865869 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-scripts\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.866722 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-ring-data-devices\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.870451 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-dispersionconf\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.871462 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-swiftconf\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.871890 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-combined-ca-bundle\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:39 crc kubenswrapper[4857]: I1201 21:53:39.893864 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhcln\" (UniqueName: \"kubernetes.io/projected/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-kube-api-access-rhcln\") pod \"swift-ring-rebalance-q5szn\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.194962 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.337077 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.357082 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.477519 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-dispersionconf\") pod \"db2d0de5-6765-4080-af32-7e98182627a1\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.477656 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db2d0de5-6765-4080-af32-7e98182627a1-etc-swift\") pod \"db2d0de5-6765-4080-af32-7e98182627a1\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.477716 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-swiftconf\") pod \"db2d0de5-6765-4080-af32-7e98182627a1\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.477752 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db2d0de5-6765-4080-af32-7e98182627a1-ring-data-devices\") pod \"db2d0de5-6765-4080-af32-7e98182627a1\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.477782 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db2d0de5-6765-4080-af32-7e98182627a1-scripts\") pod \"db2d0de5-6765-4080-af32-7e98182627a1\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.477811 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv4tz\" (UniqueName: \"kubernetes.io/projected/db2d0de5-6765-4080-af32-7e98182627a1-kube-api-access-hv4tz\") pod \"db2d0de5-6765-4080-af32-7e98182627a1\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.477843 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-combined-ca-bundle\") pod \"db2d0de5-6765-4080-af32-7e98182627a1\" (UID: \"db2d0de5-6765-4080-af32-7e98182627a1\") " Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.478716 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db2d0de5-6765-4080-af32-7e98182627a1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "db2d0de5-6765-4080-af32-7e98182627a1" (UID: "db2d0de5-6765-4080-af32-7e98182627a1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.479059 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db2d0de5-6765-4080-af32-7e98182627a1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "db2d0de5-6765-4080-af32-7e98182627a1" (UID: "db2d0de5-6765-4080-af32-7e98182627a1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.479118 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db2d0de5-6765-4080-af32-7e98182627a1-scripts" (OuterVolumeSpecName: "scripts") pod "db2d0de5-6765-4080-af32-7e98182627a1" (UID: "db2d0de5-6765-4080-af32-7e98182627a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.486220 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db2d0de5-6765-4080-af32-7e98182627a1-kube-api-access-hv4tz" (OuterVolumeSpecName: "kube-api-access-hv4tz") pod "db2d0de5-6765-4080-af32-7e98182627a1" (UID: "db2d0de5-6765-4080-af32-7e98182627a1"). InnerVolumeSpecName "kube-api-access-hv4tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.486778 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db2d0de5-6765-4080-af32-7e98182627a1" (UID: "db2d0de5-6765-4080-af32-7e98182627a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.487492 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "db2d0de5-6765-4080-af32-7e98182627a1" (UID: "db2d0de5-6765-4080-af32-7e98182627a1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.489221 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "db2d0de5-6765-4080-af32-7e98182627a1" (UID: "db2d0de5-6765-4080-af32-7e98182627a1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.581454 4857 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.581497 4857 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db2d0de5-6765-4080-af32-7e98182627a1-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.581513 4857 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.581525 4857 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db2d0de5-6765-4080-af32-7e98182627a1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.581539 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db2d0de5-6765-4080-af32-7e98182627a1-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.581551 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv4tz\" (UniqueName: \"kubernetes.io/projected/db2d0de5-6765-4080-af32-7e98182627a1-kube-api-access-hv4tz\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.581587 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2d0de5-6765-4080-af32-7e98182627a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:40 crc kubenswrapper[4857]: I1201 21:53:40.761655 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-q5szn"] Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.362167 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xptvd" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.362185 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-q5szn" event={"ID":"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b","Type":"ContainerStarted","Data":"325363d28ef0489f5d1239a9cd4c5c8a88677d7cd6769065c716943eb44cc751"} Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.377157 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8640-account-create-update-chdlg"] Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.378693 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8640-account-create-update-chdlg" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.381578 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.399845 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8640-account-create-update-chdlg"] Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.412184 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-sjprl"] Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.413192 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sjprl" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.433792 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-sjprl"] Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.443617 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-xptvd"] Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.451846 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stszp\" (UniqueName: \"kubernetes.io/projected/b06f96e5-3ee7-4b8f-95d0-968855414abe-kube-api-access-stszp\") pod \"keystone-8640-account-create-update-chdlg\" (UID: \"b06f96e5-3ee7-4b8f-95d0-968855414abe\") " pod="openstack/keystone-8640-account-create-update-chdlg" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.451905 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b06f96e5-3ee7-4b8f-95d0-968855414abe-operator-scripts\") pod \"keystone-8640-account-create-update-chdlg\" (UID: \"b06f96e5-3ee7-4b8f-95d0-968855414abe\") " pod="openstack/keystone-8640-account-create-update-chdlg" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.452024 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1dcab1d-428a-402a-a901-4d0d07464487-operator-scripts\") pod \"keystone-db-create-sjprl\" (UID: \"e1dcab1d-428a-402a-a901-4d0d07464487\") " pod="openstack/keystone-db-create-sjprl" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.452079 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7bt8\" (UniqueName: \"kubernetes.io/projected/e1dcab1d-428a-402a-a901-4d0d07464487-kube-api-access-b7bt8\") pod \"keystone-db-create-sjprl\" (UID: \"e1dcab1d-428a-402a-a901-4d0d07464487\") " pod="openstack/keystone-db-create-sjprl" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.454143 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-xptvd"] Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.553444 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stszp\" (UniqueName: \"kubernetes.io/projected/b06f96e5-3ee7-4b8f-95d0-968855414abe-kube-api-access-stszp\") pod \"keystone-8640-account-create-update-chdlg\" (UID: \"b06f96e5-3ee7-4b8f-95d0-968855414abe\") " pod="openstack/keystone-8640-account-create-update-chdlg" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.553532 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b06f96e5-3ee7-4b8f-95d0-968855414abe-operator-scripts\") pod \"keystone-8640-account-create-update-chdlg\" (UID: \"b06f96e5-3ee7-4b8f-95d0-968855414abe\") " pod="openstack/keystone-8640-account-create-update-chdlg" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.553572 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1dcab1d-428a-402a-a901-4d0d07464487-operator-scripts\") pod \"keystone-db-create-sjprl\" (UID: \"e1dcab1d-428a-402a-a901-4d0d07464487\") " pod="openstack/keystone-db-create-sjprl" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.553594 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7bt8\" (UniqueName: \"kubernetes.io/projected/e1dcab1d-428a-402a-a901-4d0d07464487-kube-api-access-b7bt8\") pod \"keystone-db-create-sjprl\" (UID: \"e1dcab1d-428a-402a-a901-4d0d07464487\") " pod="openstack/keystone-db-create-sjprl" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.554641 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b06f96e5-3ee7-4b8f-95d0-968855414abe-operator-scripts\") pod \"keystone-8640-account-create-update-chdlg\" (UID: \"b06f96e5-3ee7-4b8f-95d0-968855414abe\") " pod="openstack/keystone-8640-account-create-update-chdlg" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.555144 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1dcab1d-428a-402a-a901-4d0d07464487-operator-scripts\") pod \"keystone-db-create-sjprl\" (UID: \"e1dcab1d-428a-402a-a901-4d0d07464487\") " pod="openstack/keystone-db-create-sjprl" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.575922 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stszp\" (UniqueName: \"kubernetes.io/projected/b06f96e5-3ee7-4b8f-95d0-968855414abe-kube-api-access-stszp\") pod \"keystone-8640-account-create-update-chdlg\" (UID: \"b06f96e5-3ee7-4b8f-95d0-968855414abe\") " pod="openstack/keystone-8640-account-create-update-chdlg" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.575967 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7bt8\" (UniqueName: \"kubernetes.io/projected/e1dcab1d-428a-402a-a901-4d0d07464487-kube-api-access-b7bt8\") pod \"keystone-db-create-sjprl\" (UID: \"e1dcab1d-428a-402a-a901-4d0d07464487\") " pod="openstack/keystone-db-create-sjprl" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.640013 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-qlvz9"] Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.641300 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qlvz9" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.651130 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-qlvz9"] Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.704271 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8640-account-create-update-chdlg" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.733532 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sjprl" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.757733 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f-operator-scripts\") pod \"placement-db-create-qlvz9\" (UID: \"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f\") " pod="openstack/placement-db-create-qlvz9" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.757877 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hmkd\" (UniqueName: \"kubernetes.io/projected/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f-kube-api-access-4hmkd\") pod \"placement-db-create-qlvz9\" (UID: \"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f\") " pod="openstack/placement-db-create-qlvz9" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.804376 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-75e9-account-create-update-lqsf5"] Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.806755 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75e9-account-create-update-lqsf5" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.810197 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.812677 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-75e9-account-create-update-lqsf5"] Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.877904 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hmkd\" (UniqueName: \"kubernetes.io/projected/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f-kube-api-access-4hmkd\") pod \"placement-db-create-qlvz9\" (UID: \"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f\") " pod="openstack/placement-db-create-qlvz9" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.883419 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f-operator-scripts\") pod \"placement-db-create-qlvz9\" (UID: \"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f\") " pod="openstack/placement-db-create-qlvz9" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.887678 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f-operator-scripts\") pod \"placement-db-create-qlvz9\" (UID: \"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f\") " pod="openstack/placement-db-create-qlvz9" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.892124 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db2d0de5-6765-4080-af32-7e98182627a1" path="/var/lib/kubelet/pods/db2d0de5-6765-4080-af32-7e98182627a1/volumes" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.921548 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hmkd\" (UniqueName: \"kubernetes.io/projected/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f-kube-api-access-4hmkd\") pod \"placement-db-create-qlvz9\" (UID: \"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f\") " pod="openstack/placement-db-create-qlvz9" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.987348 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31eea916-82d4-4918-8577-68276ba53766-operator-scripts\") pod \"placement-75e9-account-create-update-lqsf5\" (UID: \"31eea916-82d4-4918-8577-68276ba53766\") " pod="openstack/placement-75e9-account-create-update-lqsf5" Dec 01 21:53:41 crc kubenswrapper[4857]: I1201 21:53:41.987499 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk95n\" (UniqueName: \"kubernetes.io/projected/31eea916-82d4-4918-8577-68276ba53766-kube-api-access-zk95n\") pod \"placement-75e9-account-create-update-lqsf5\" (UID: \"31eea916-82d4-4918-8577-68276ba53766\") " pod="openstack/placement-75e9-account-create-update-lqsf5" Dec 01 21:53:42 crc kubenswrapper[4857]: I1201 21:53:42.014524 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qlvz9" Dec 01 21:53:42 crc kubenswrapper[4857]: I1201 21:53:42.166677 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk95n\" (UniqueName: \"kubernetes.io/projected/31eea916-82d4-4918-8577-68276ba53766-kube-api-access-zk95n\") pod \"placement-75e9-account-create-update-lqsf5\" (UID: \"31eea916-82d4-4918-8577-68276ba53766\") " pod="openstack/placement-75e9-account-create-update-lqsf5" Dec 01 21:53:42 crc kubenswrapper[4857]: I1201 21:53:42.166775 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31eea916-82d4-4918-8577-68276ba53766-operator-scripts\") pod \"placement-75e9-account-create-update-lqsf5\" (UID: \"31eea916-82d4-4918-8577-68276ba53766\") " pod="openstack/placement-75e9-account-create-update-lqsf5" Dec 01 21:53:42 crc kubenswrapper[4857]: I1201 21:53:42.167927 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31eea916-82d4-4918-8577-68276ba53766-operator-scripts\") pod \"placement-75e9-account-create-update-lqsf5\" (UID: \"31eea916-82d4-4918-8577-68276ba53766\") " pod="openstack/placement-75e9-account-create-update-lqsf5" Dec 01 21:53:42 crc kubenswrapper[4857]: I1201 21:53:42.217817 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk95n\" (UniqueName: \"kubernetes.io/projected/31eea916-82d4-4918-8577-68276ba53766-kube-api-access-zk95n\") pod \"placement-75e9-account-create-update-lqsf5\" (UID: \"31eea916-82d4-4918-8577-68276ba53766\") " pod="openstack/placement-75e9-account-create-update-lqsf5" Dec 01 21:53:42 crc kubenswrapper[4857]: I1201 21:53:42.270411 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-sjprl"] Dec 01 21:53:42 crc kubenswrapper[4857]: I1201 21:53:42.380378 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sjprl" event={"ID":"e1dcab1d-428a-402a-a901-4d0d07464487","Type":"ContainerStarted","Data":"2fcc47c50033b38c2ff800efd856394b406893c29de1944849454ace99905559"} Dec 01 21:53:42 crc kubenswrapper[4857]: I1201 21:53:42.475531 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75e9-account-create-update-lqsf5" Dec 01 21:53:42 crc kubenswrapper[4857]: I1201 21:53:42.510498 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8640-account-create-update-chdlg"] Dec 01 21:53:42 crc kubenswrapper[4857]: W1201 21:53:42.521884 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb06f96e5_3ee7_4b8f_95d0_968855414abe.slice/crio-1dd0dafcbbe457eed77b0a69d0e0dd9009bc26484fa7c312c22752975a73bf47 WatchSource:0}: Error finding container 1dd0dafcbbe457eed77b0a69d0e0dd9009bc26484fa7c312c22752975a73bf47: Status 404 returned error can't find the container with id 1dd0dafcbbe457eed77b0a69d0e0dd9009bc26484fa7c312c22752975a73bf47 Dec 01 21:53:42 crc kubenswrapper[4857]: I1201 21:53:42.529801 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 01 21:53:42 crc kubenswrapper[4857]: I1201 21:53:42.696383 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-qlvz9"] Dec 01 21:53:42 crc kubenswrapper[4857]: I1201 21:53:42.921124 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-75e9-account-create-update-lqsf5"] Dec 01 21:53:43 crc kubenswrapper[4857]: I1201 21:53:43.196408 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:43 crc kubenswrapper[4857]: E1201 21:53:43.196602 4857 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 21:53:43 crc kubenswrapper[4857]: E1201 21:53:43.196631 4857 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 21:53:43 crc kubenswrapper[4857]: E1201 21:53:43.196708 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift podName:3ea80def-31e0-4cc2-9f15-26ee35c62641 nodeName:}" failed. No retries permitted until 2025-12-01 21:53:51.196684029 +0000 UTC m=+1149.686746366 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift") pod "swift-storage-0" (UID: "3ea80def-31e0-4cc2-9f15-26ee35c62641") : configmap "swift-ring-files" not found Dec 01 21:53:43 crc kubenswrapper[4857]: I1201 21:53:43.392464 4857 generic.go:334] "Generic (PLEG): container finished" podID="b06f96e5-3ee7-4b8f-95d0-968855414abe" containerID="0c0f19e0917b3adb587ed1cee632a48dc1f82d6382e639315247141964c03f6b" exitCode=0 Dec 01 21:53:43 crc kubenswrapper[4857]: I1201 21:53:43.392559 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8640-account-create-update-chdlg" event={"ID":"b06f96e5-3ee7-4b8f-95d0-968855414abe","Type":"ContainerDied","Data":"0c0f19e0917b3adb587ed1cee632a48dc1f82d6382e639315247141964c03f6b"} Dec 01 21:53:43 crc kubenswrapper[4857]: I1201 21:53:43.392593 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8640-account-create-update-chdlg" event={"ID":"b06f96e5-3ee7-4b8f-95d0-968855414abe","Type":"ContainerStarted","Data":"1dd0dafcbbe457eed77b0a69d0e0dd9009bc26484fa7c312c22752975a73bf47"} Dec 01 21:53:43 crc kubenswrapper[4857]: I1201 21:53:43.395114 4857 generic.go:334] "Generic (PLEG): container finished" podID="e1dcab1d-428a-402a-a901-4d0d07464487" containerID="4306ceb75c17d7dafa76df7b3e43f7eaf6d2b959b17e9aaca62e1e773f1b76a6" exitCode=0 Dec 01 21:53:43 crc kubenswrapper[4857]: I1201 21:53:43.395164 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sjprl" event={"ID":"e1dcab1d-428a-402a-a901-4d0d07464487","Type":"ContainerDied","Data":"4306ceb75c17d7dafa76df7b3e43f7eaf6d2b959b17e9aaca62e1e773f1b76a6"} Dec 01 21:53:44 crc kubenswrapper[4857]: I1201 21:53:44.569036 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:53:44 crc kubenswrapper[4857]: I1201 21:53:44.650141 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-g6z4j"] Dec 01 21:53:44 crc kubenswrapper[4857]: I1201 21:53:44.650553 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" podUID="8e724c46-b6d3-4e20-910b-8a7e417fd0cb" containerName="dnsmasq-dns" containerID="cri-o://d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445" gracePeriod=10 Dec 01 21:53:44 crc kubenswrapper[4857]: I1201 21:53:44.929063 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sjprl" Dec 01 21:53:44 crc kubenswrapper[4857]: W1201 21:53:44.942848 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31eea916_82d4_4918_8577_68276ba53766.slice/crio-6b2420b0f8639c51436f888a97ff6d39a0df0e4ebeb0e77982b9c323444d8466 WatchSource:0}: Error finding container 6b2420b0f8639c51436f888a97ff6d39a0df0e4ebeb0e77982b9c323444d8466: Status 404 returned error can't find the container with id 6b2420b0f8639c51436f888a97ff6d39a0df0e4ebeb0e77982b9c323444d8466 Dec 01 21:53:44 crc kubenswrapper[4857]: I1201 21:53:44.975865 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.034718 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1dcab1d-428a-402a-a901-4d0d07464487-operator-scripts\") pod \"e1dcab1d-428a-402a-a901-4d0d07464487\" (UID: \"e1dcab1d-428a-402a-a901-4d0d07464487\") " Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.035061 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7bt8\" (UniqueName: \"kubernetes.io/projected/e1dcab1d-428a-402a-a901-4d0d07464487-kube-api-access-b7bt8\") pod \"e1dcab1d-428a-402a-a901-4d0d07464487\" (UID: \"e1dcab1d-428a-402a-a901-4d0d07464487\") " Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.037280 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1dcab1d-428a-402a-a901-4d0d07464487-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e1dcab1d-428a-402a-a901-4d0d07464487" (UID: "e1dcab1d-428a-402a-a901-4d0d07464487"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.039271 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1dcab1d-428a-402a-a901-4d0d07464487-kube-api-access-b7bt8" (OuterVolumeSpecName: "kube-api-access-b7bt8") pod "e1dcab1d-428a-402a-a901-4d0d07464487" (UID: "e1dcab1d-428a-402a-a901-4d0d07464487"). InnerVolumeSpecName "kube-api-access-b7bt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.092974 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8640-account-create-update-chdlg" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.137009 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b06f96e5-3ee7-4b8f-95d0-968855414abe-operator-scripts\") pod \"b06f96e5-3ee7-4b8f-95d0-968855414abe\" (UID: \"b06f96e5-3ee7-4b8f-95d0-968855414abe\") " Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.137162 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stszp\" (UniqueName: \"kubernetes.io/projected/b06f96e5-3ee7-4b8f-95d0-968855414abe-kube-api-access-stszp\") pod \"b06f96e5-3ee7-4b8f-95d0-968855414abe\" (UID: \"b06f96e5-3ee7-4b8f-95d0-968855414abe\") " Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.137533 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1dcab1d-428a-402a-a901-4d0d07464487-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.137551 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7bt8\" (UniqueName: \"kubernetes.io/projected/e1dcab1d-428a-402a-a901-4d0d07464487-kube-api-access-b7bt8\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.138130 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b06f96e5-3ee7-4b8f-95d0-968855414abe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b06f96e5-3ee7-4b8f-95d0-968855414abe" (UID: "b06f96e5-3ee7-4b8f-95d0-968855414abe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.146408 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b06f96e5-3ee7-4b8f-95d0-968855414abe-kube-api-access-stszp" (OuterVolumeSpecName: "kube-api-access-stszp") pod "b06f96e5-3ee7-4b8f-95d0-968855414abe" (UID: "b06f96e5-3ee7-4b8f-95d0-968855414abe"). InnerVolumeSpecName "kube-api-access-stszp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.244470 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stszp\" (UniqueName: \"kubernetes.io/projected/b06f96e5-3ee7-4b8f-95d0-968855414abe-kube-api-access-stszp\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.244517 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b06f96e5-3ee7-4b8f-95d0-968855414abe-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.260263 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.346315 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2smx7\" (UniqueName: \"kubernetes.io/projected/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-kube-api-access-2smx7\") pod \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.346374 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-ovsdbserver-nb\") pod \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.346474 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-ovsdbserver-sb\") pod \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.346548 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-dns-svc\") pod \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.346574 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-config\") pod \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\" (UID: \"8e724c46-b6d3-4e20-910b-8a7e417fd0cb\") " Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.353222 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-kube-api-access-2smx7" (OuterVolumeSpecName: "kube-api-access-2smx7") pod "8e724c46-b6d3-4e20-910b-8a7e417fd0cb" (UID: "8e724c46-b6d3-4e20-910b-8a7e417fd0cb"). InnerVolumeSpecName "kube-api-access-2smx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.391749 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e724c46-b6d3-4e20-910b-8a7e417fd0cb" (UID: "8e724c46-b6d3-4e20-910b-8a7e417fd0cb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.394171 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e724c46-b6d3-4e20-910b-8a7e417fd0cb" (UID: "8e724c46-b6d3-4e20-910b-8a7e417fd0cb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.395116 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-config" (OuterVolumeSpecName: "config") pod "8e724c46-b6d3-4e20-910b-8a7e417fd0cb" (UID: "8e724c46-b6d3-4e20-910b-8a7e417fd0cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.396429 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e724c46-b6d3-4e20-910b-8a7e417fd0cb" (UID: "8e724c46-b6d3-4e20-910b-8a7e417fd0cb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.417531 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75e9-account-create-update-lqsf5" event={"ID":"31eea916-82d4-4918-8577-68276ba53766","Type":"ContainerStarted","Data":"9a93c47608aaad7c765c2674da01cc28398fbd69452fd420bf3fd191c83b13fc"} Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.417614 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75e9-account-create-update-lqsf5" event={"ID":"31eea916-82d4-4918-8577-68276ba53766","Type":"ContainerStarted","Data":"6b2420b0f8639c51436f888a97ff6d39a0df0e4ebeb0e77982b9c323444d8466"} Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.420813 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-q5szn" event={"ID":"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b","Type":"ContainerStarted","Data":"476ec1f85af117a366e28b21da443ea58ff745bcf0964eddef9c6cdf44627d43"} Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.422870 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8640-account-create-update-chdlg" event={"ID":"b06f96e5-3ee7-4b8f-95d0-968855414abe","Type":"ContainerDied","Data":"1dd0dafcbbe457eed77b0a69d0e0dd9009bc26484fa7c312c22752975a73bf47"} Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.422915 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dd0dafcbbe457eed77b0a69d0e0dd9009bc26484fa7c312c22752975a73bf47" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.422969 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8640-account-create-update-chdlg" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.427481 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sjprl" event={"ID":"e1dcab1d-428a-402a-a901-4d0d07464487","Type":"ContainerDied","Data":"2fcc47c50033b38c2ff800efd856394b406893c29de1944849454ace99905559"} Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.427507 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fcc47c50033b38c2ff800efd856394b406893c29de1944849454ace99905559" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.427565 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sjprl" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.434314 4857 generic.go:334] "Generic (PLEG): container finished" podID="5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f" containerID="ca81184d8dca1b486457e141f51b17710fc9c8e1b9b3d57ea5b58971f2fc3a6b" exitCode=0 Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.434451 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qlvz9" event={"ID":"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f","Type":"ContainerDied","Data":"ca81184d8dca1b486457e141f51b17710fc9c8e1b9b3d57ea5b58971f2fc3a6b"} Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.434862 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qlvz9" event={"ID":"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f","Type":"ContainerStarted","Data":"cbc6e438e6a28b02cac86f5e4ec26750f9b0a65d52572e93f5d2d2edb9cc7cc9"} Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.438326 4857 generic.go:334] "Generic (PLEG): container finished" podID="8e724c46-b6d3-4e20-910b-8a7e417fd0cb" containerID="d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445" exitCode=0 Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.438391 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" event={"ID":"8e724c46-b6d3-4e20-910b-8a7e417fd0cb","Type":"ContainerDied","Data":"d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445"} Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.438435 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" event={"ID":"8e724c46-b6d3-4e20-910b-8a7e417fd0cb","Type":"ContainerDied","Data":"4845bf4381cfcdcf9db89c765f6fca6219a61c2db671b8f3abb1b9e1551d49a6"} Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.438459 4857 scope.go:117] "RemoveContainer" containerID="d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.438634 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-g6z4j" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.445394 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-75e9-account-create-update-lqsf5" podStartSLOduration=4.445364741 podStartE2EDuration="4.445364741s" podCreationTimestamp="2025-12-01 21:53:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:53:45.4358454 +0000 UTC m=+1143.925907717" watchObservedRunningTime="2025-12-01 21:53:45.445364741 +0000 UTC m=+1143.935427058" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.448659 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.448687 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.448701 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.448717 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2smx7\" (UniqueName: \"kubernetes.io/projected/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-kube-api-access-2smx7\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.448731 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e724c46-b6d3-4e20-910b-8a7e417fd0cb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.464183 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-q5szn" podStartSLOduration=2.227781602 podStartE2EDuration="6.464164056s" podCreationTimestamp="2025-12-01 21:53:39 +0000 UTC" firstStartedPulling="2025-12-01 21:53:40.775440859 +0000 UTC m=+1139.265503216" lastFinishedPulling="2025-12-01 21:53:45.011823353 +0000 UTC m=+1143.501885670" observedRunningTime="2025-12-01 21:53:45.453008046 +0000 UTC m=+1143.943070363" watchObservedRunningTime="2025-12-01 21:53:45.464164056 +0000 UTC m=+1143.954226373" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.550287 4857 scope.go:117] "RemoveContainer" containerID="fa02bc89e5d41b2cf8e97787bed6fedf04aafe584d76281cdf7495bdf95ef3fc" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.567236 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-g6z4j"] Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.574832 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-g6z4j"] Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.600320 4857 scope.go:117] "RemoveContainer" containerID="d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445" Dec 01 21:53:45 crc kubenswrapper[4857]: E1201 21:53:45.603985 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445\": container with ID starting with d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445 not found: ID does not exist" containerID="d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.604031 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445"} err="failed to get container status \"d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445\": rpc error: code = NotFound desc = could not find container \"d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445\": container with ID starting with d2ef048003e609149a3a434744702490df46c8682bb4388cf8d26e48a8bf6445 not found: ID does not exist" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.604076 4857 scope.go:117] "RemoveContainer" containerID="fa02bc89e5d41b2cf8e97787bed6fedf04aafe584d76281cdf7495bdf95ef3fc" Dec 01 21:53:45 crc kubenswrapper[4857]: E1201 21:53:45.604527 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa02bc89e5d41b2cf8e97787bed6fedf04aafe584d76281cdf7495bdf95ef3fc\": container with ID starting with fa02bc89e5d41b2cf8e97787bed6fedf04aafe584d76281cdf7495bdf95ef3fc not found: ID does not exist" containerID="fa02bc89e5d41b2cf8e97787bed6fedf04aafe584d76281cdf7495bdf95ef3fc" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.604587 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa02bc89e5d41b2cf8e97787bed6fedf04aafe584d76281cdf7495bdf95ef3fc"} err="failed to get container status \"fa02bc89e5d41b2cf8e97787bed6fedf04aafe584d76281cdf7495bdf95ef3fc\": rpc error: code = NotFound desc = could not find container \"fa02bc89e5d41b2cf8e97787bed6fedf04aafe584d76281cdf7495bdf95ef3fc\": container with ID starting with fa02bc89e5d41b2cf8e97787bed6fedf04aafe584d76281cdf7495bdf95ef3fc not found: ID does not exist" Dec 01 21:53:45 crc kubenswrapper[4857]: I1201 21:53:45.846772 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e724c46-b6d3-4e20-910b-8a7e417fd0cb" path="/var/lib/kubelet/pods/8e724c46-b6d3-4e20-910b-8a7e417fd0cb/volumes" Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.449890 4857 generic.go:334] "Generic (PLEG): container finished" podID="84a09f7d-b0a7-4055-93e1-83d11edb6467" containerID="8e89121786dac33726d523e59f2c80de2165f73812834f8643c69bd32e7e2d1c" exitCode=0 Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.449906 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"84a09f7d-b0a7-4055-93e1-83d11edb6467","Type":"ContainerDied","Data":"8e89121786dac33726d523e59f2c80de2165f73812834f8643c69bd32e7e2d1c"} Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.452428 4857 generic.go:334] "Generic (PLEG): container finished" podID="235e0415-1e77-480d-b3da-157ed0252d4f" containerID="868e89c708562210a5a6e915928b232c0e7618e25b2d87a0152f36a0bcbc14ad" exitCode=0 Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.452497 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"235e0415-1e77-480d-b3da-157ed0252d4f","Type":"ContainerDied","Data":"868e89c708562210a5a6e915928b232c0e7618e25b2d87a0152f36a0bcbc14ad"} Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.456378 4857 generic.go:334] "Generic (PLEG): container finished" podID="31eea916-82d4-4918-8577-68276ba53766" containerID="9a93c47608aaad7c765c2674da01cc28398fbd69452fd420bf3fd191c83b13fc" exitCode=0 Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.456665 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75e9-account-create-update-lqsf5" event={"ID":"31eea916-82d4-4918-8577-68276ba53766","Type":"ContainerDied","Data":"9a93c47608aaad7c765c2674da01cc28398fbd69452fd420bf3fd191c83b13fc"} Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.811274 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qlvz9" Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.955695 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-5zzh6"] Dec 01 21:53:46 crc kubenswrapper[4857]: E1201 21:53:46.956095 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b06f96e5-3ee7-4b8f-95d0-968855414abe" containerName="mariadb-account-create-update" Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.956111 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b06f96e5-3ee7-4b8f-95d0-968855414abe" containerName="mariadb-account-create-update" Dec 01 21:53:46 crc kubenswrapper[4857]: E1201 21:53:46.956129 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f" containerName="mariadb-database-create" Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.956140 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f" containerName="mariadb-database-create" Dec 01 21:53:46 crc kubenswrapper[4857]: E1201 21:53:46.956152 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1dcab1d-428a-402a-a901-4d0d07464487" containerName="mariadb-database-create" Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.956158 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1dcab1d-428a-402a-a901-4d0d07464487" containerName="mariadb-database-create" Dec 01 21:53:46 crc kubenswrapper[4857]: E1201 21:53:46.956179 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e724c46-b6d3-4e20-910b-8a7e417fd0cb" containerName="init" Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.956185 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e724c46-b6d3-4e20-910b-8a7e417fd0cb" containerName="init" Dec 01 21:53:46 crc kubenswrapper[4857]: E1201 21:53:46.956200 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e724c46-b6d3-4e20-910b-8a7e417fd0cb" containerName="dnsmasq-dns" Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.956205 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e724c46-b6d3-4e20-910b-8a7e417fd0cb" containerName="dnsmasq-dns" Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.957690 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1dcab1d-428a-402a-a901-4d0d07464487" containerName="mariadb-database-create" Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.957824 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b06f96e5-3ee7-4b8f-95d0-968855414abe" containerName="mariadb-account-create-update" Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.957868 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e724c46-b6d3-4e20-910b-8a7e417fd0cb" containerName="dnsmasq-dns" Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.957885 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f" containerName="mariadb-database-create" Dec 01 21:53:46 crc kubenswrapper[4857]: I1201 21:53:46.959340 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5zzh6" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:46.987113 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f-operator-scripts\") pod \"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f\" (UID: \"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f\") " Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:46.987216 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hmkd\" (UniqueName: \"kubernetes.io/projected/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f-kube-api-access-4hmkd\") pod \"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f\" (UID: \"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f\") " Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:46.992650 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f" (UID: "5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.008086 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5zzh6"] Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.012195 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f-kube-api-access-4hmkd" (OuterVolumeSpecName: "kube-api-access-4hmkd") pod "5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f" (UID: "5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f"). InnerVolumeSpecName "kube-api-access-4hmkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.022130 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-41ef-account-create-update-rfn85"] Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.031013 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-41ef-account-create-update-rfn85" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.034632 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.060906 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-41ef-account-create-update-rfn85"] Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.090153 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt6dl\" (UniqueName: \"kubernetes.io/projected/55e4213d-69d8-4f02-adcb-e2249fb3dccf-kube-api-access-dt6dl\") pod \"glance-db-create-5zzh6\" (UID: \"55e4213d-69d8-4f02-adcb-e2249fb3dccf\") " pod="openstack/glance-db-create-5zzh6" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.090252 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55e4213d-69d8-4f02-adcb-e2249fb3dccf-operator-scripts\") pod \"glance-db-create-5zzh6\" (UID: \"55e4213d-69d8-4f02-adcb-e2249fb3dccf\") " pod="openstack/glance-db-create-5zzh6" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.090397 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.090411 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hmkd\" (UniqueName: \"kubernetes.io/projected/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f-kube-api-access-4hmkd\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.191757 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6fbba64-3540-42d5-a1dd-c5e65d4e1330-operator-scripts\") pod \"glance-41ef-account-create-update-rfn85\" (UID: \"b6fbba64-3540-42d5-a1dd-c5e65d4e1330\") " pod="openstack/glance-41ef-account-create-update-rfn85" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.191809 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55e4213d-69d8-4f02-adcb-e2249fb3dccf-operator-scripts\") pod \"glance-db-create-5zzh6\" (UID: \"55e4213d-69d8-4f02-adcb-e2249fb3dccf\") " pod="openstack/glance-db-create-5zzh6" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.191904 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47h69\" (UniqueName: \"kubernetes.io/projected/b6fbba64-3540-42d5-a1dd-c5e65d4e1330-kube-api-access-47h69\") pod \"glance-41ef-account-create-update-rfn85\" (UID: \"b6fbba64-3540-42d5-a1dd-c5e65d4e1330\") " pod="openstack/glance-41ef-account-create-update-rfn85" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.191947 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt6dl\" (UniqueName: \"kubernetes.io/projected/55e4213d-69d8-4f02-adcb-e2249fb3dccf-kube-api-access-dt6dl\") pod \"glance-db-create-5zzh6\" (UID: \"55e4213d-69d8-4f02-adcb-e2249fb3dccf\") " pod="openstack/glance-db-create-5zzh6" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.192697 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55e4213d-69d8-4f02-adcb-e2249fb3dccf-operator-scripts\") pod \"glance-db-create-5zzh6\" (UID: \"55e4213d-69d8-4f02-adcb-e2249fb3dccf\") " pod="openstack/glance-db-create-5zzh6" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.211953 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt6dl\" (UniqueName: \"kubernetes.io/projected/55e4213d-69d8-4f02-adcb-e2249fb3dccf-kube-api-access-dt6dl\") pod \"glance-db-create-5zzh6\" (UID: \"55e4213d-69d8-4f02-adcb-e2249fb3dccf\") " pod="openstack/glance-db-create-5zzh6" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.292926 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47h69\" (UniqueName: \"kubernetes.io/projected/b6fbba64-3540-42d5-a1dd-c5e65d4e1330-kube-api-access-47h69\") pod \"glance-41ef-account-create-update-rfn85\" (UID: \"b6fbba64-3540-42d5-a1dd-c5e65d4e1330\") " pod="openstack/glance-41ef-account-create-update-rfn85" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.293342 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6fbba64-3540-42d5-a1dd-c5e65d4e1330-operator-scripts\") pod \"glance-41ef-account-create-update-rfn85\" (UID: \"b6fbba64-3540-42d5-a1dd-c5e65d4e1330\") " pod="openstack/glance-41ef-account-create-update-rfn85" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.294070 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6fbba64-3540-42d5-a1dd-c5e65d4e1330-operator-scripts\") pod \"glance-41ef-account-create-update-rfn85\" (UID: \"b6fbba64-3540-42d5-a1dd-c5e65d4e1330\") " pod="openstack/glance-41ef-account-create-update-rfn85" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.312800 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47h69\" (UniqueName: \"kubernetes.io/projected/b6fbba64-3540-42d5-a1dd-c5e65d4e1330-kube-api-access-47h69\") pod \"glance-41ef-account-create-update-rfn85\" (UID: \"b6fbba64-3540-42d5-a1dd-c5e65d4e1330\") " pod="openstack/glance-41ef-account-create-update-rfn85" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.346760 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5zzh6" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.362720 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-41ef-account-create-update-rfn85" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.469340 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"84a09f7d-b0a7-4055-93e1-83d11edb6467","Type":"ContainerStarted","Data":"55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb"} Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.470394 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.472909 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qlvz9" event={"ID":"5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f","Type":"ContainerDied","Data":"cbc6e438e6a28b02cac86f5e4ec26750f9b0a65d52572e93f5d2d2edb9cc7cc9"} Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.472934 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbc6e438e6a28b02cac86f5e4ec26750f9b0a65d52572e93f5d2d2edb9cc7cc9" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.472978 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qlvz9" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.475626 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nhfhn" podUID="32213ad3-4287-4968-9d42-bfeb2ff558d1" containerName="ovn-controller" probeResult="failure" output=< Dec 01 21:53:47 crc kubenswrapper[4857]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 21:53:47 crc kubenswrapper[4857]: > Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.477136 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"235e0415-1e77-480d-b3da-157ed0252d4f","Type":"ContainerStarted","Data":"61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce"} Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.477365 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.520673 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.153258083 podStartE2EDuration="1m10.520653184s" podCreationTimestamp="2025-12-01 21:52:37 +0000 UTC" firstStartedPulling="2025-12-01 21:52:40.517103794 +0000 UTC m=+1079.007166111" lastFinishedPulling="2025-12-01 21:53:11.884498855 +0000 UTC m=+1110.374561212" observedRunningTime="2025-12-01 21:53:47.49776592 +0000 UTC m=+1145.987828237" watchObservedRunningTime="2025-12-01 21:53:47.520653184 +0000 UTC m=+1146.010715501" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.537073 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.129894238 podStartE2EDuration="1m11.537056651s" podCreationTimestamp="2025-12-01 21:52:36 +0000 UTC" firstStartedPulling="2025-12-01 21:52:39.57178557 +0000 UTC m=+1078.061847887" lastFinishedPulling="2025-12-01 21:53:11.978947983 +0000 UTC m=+1110.469010300" observedRunningTime="2025-12-01 21:53:47.532516351 +0000 UTC m=+1146.022578668" watchObservedRunningTime="2025-12-01 21:53:47.537056651 +0000 UTC m=+1146.027118958" Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.845591 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-41ef-account-create-update-rfn85"] Dec 01 21:53:47 crc kubenswrapper[4857]: I1201 21:53:47.925579 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5zzh6"] Dec 01 21:53:47 crc kubenswrapper[4857]: W1201 21:53:47.938027 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55e4213d_69d8_4f02_adcb_e2249fb3dccf.slice/crio-54718abecf0dcd430b9af7ca623a8913d80dd1406d615f6e4f6b9a869fc1e0af WatchSource:0}: Error finding container 54718abecf0dcd430b9af7ca623a8913d80dd1406d615f6e4f6b9a869fc1e0af: Status 404 returned error can't find the container with id 54718abecf0dcd430b9af7ca623a8913d80dd1406d615f6e4f6b9a869fc1e0af Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.113909 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75e9-account-create-update-lqsf5" Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.216358 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31eea916-82d4-4918-8577-68276ba53766-operator-scripts\") pod \"31eea916-82d4-4918-8577-68276ba53766\" (UID: \"31eea916-82d4-4918-8577-68276ba53766\") " Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.216531 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zk95n\" (UniqueName: \"kubernetes.io/projected/31eea916-82d4-4918-8577-68276ba53766-kube-api-access-zk95n\") pod \"31eea916-82d4-4918-8577-68276ba53766\" (UID: \"31eea916-82d4-4918-8577-68276ba53766\") " Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.218409 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31eea916-82d4-4918-8577-68276ba53766-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "31eea916-82d4-4918-8577-68276ba53766" (UID: "31eea916-82d4-4918-8577-68276ba53766"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.226485 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31eea916-82d4-4918-8577-68276ba53766-kube-api-access-zk95n" (OuterVolumeSpecName: "kube-api-access-zk95n") pod "31eea916-82d4-4918-8577-68276ba53766" (UID: "31eea916-82d4-4918-8577-68276ba53766"). InnerVolumeSpecName "kube-api-access-zk95n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.318182 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31eea916-82d4-4918-8577-68276ba53766-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.318587 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zk95n\" (UniqueName: \"kubernetes.io/projected/31eea916-82d4-4918-8577-68276ba53766-kube-api-access-zk95n\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.487657 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5zzh6" event={"ID":"55e4213d-69d8-4f02-adcb-e2249fb3dccf","Type":"ContainerStarted","Data":"2361a36edb8bf39369dbc352f577e586f8366a33a92ad0d587cced04acc51a77"} Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.487736 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5zzh6" event={"ID":"55e4213d-69d8-4f02-adcb-e2249fb3dccf","Type":"ContainerStarted","Data":"54718abecf0dcd430b9af7ca623a8913d80dd1406d615f6e4f6b9a869fc1e0af"} Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.491241 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-41ef-account-create-update-rfn85" event={"ID":"b6fbba64-3540-42d5-a1dd-c5e65d4e1330","Type":"ContainerStarted","Data":"2b6870dfc6a6aa4776bad120ab2416e26e4267b1479e1694092999e8e19d0a5d"} Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.491430 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-41ef-account-create-update-rfn85" event={"ID":"b6fbba64-3540-42d5-a1dd-c5e65d4e1330","Type":"ContainerStarted","Data":"319f4374adefbf6af796b214a2d291c12a3496571d0ce618f667f73d40c6276d"} Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.493620 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75e9-account-create-update-lqsf5" event={"ID":"31eea916-82d4-4918-8577-68276ba53766","Type":"ContainerDied","Data":"6b2420b0f8639c51436f888a97ff6d39a0df0e4ebeb0e77982b9c323444d8466"} Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.493673 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b2420b0f8639c51436f888a97ff6d39a0df0e4ebeb0e77982b9c323444d8466" Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.493646 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75e9-account-create-update-lqsf5" Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.516280 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-5zzh6" podStartSLOduration=2.516258983 podStartE2EDuration="2.516258983s" podCreationTimestamp="2025-12-01 21:53:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:53:48.516029877 +0000 UTC m=+1147.006092194" watchObservedRunningTime="2025-12-01 21:53:48.516258983 +0000 UTC m=+1147.006321300" Dec 01 21:53:48 crc kubenswrapper[4857]: I1201 21:53:48.533425 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-41ef-account-create-update-rfn85" podStartSLOduration=2.5333998380000002 podStartE2EDuration="2.533399838s" podCreationTimestamp="2025-12-01 21:53:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:53:48.531201635 +0000 UTC m=+1147.021263952" watchObservedRunningTime="2025-12-01 21:53:48.533399838 +0000 UTC m=+1147.023462155" Dec 01 21:53:49 crc kubenswrapper[4857]: I1201 21:53:49.506613 4857 generic.go:334] "Generic (PLEG): container finished" podID="b6fbba64-3540-42d5-a1dd-c5e65d4e1330" containerID="2b6870dfc6a6aa4776bad120ab2416e26e4267b1479e1694092999e8e19d0a5d" exitCode=0 Dec 01 21:53:49 crc kubenswrapper[4857]: I1201 21:53:49.506708 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-41ef-account-create-update-rfn85" event={"ID":"b6fbba64-3540-42d5-a1dd-c5e65d4e1330","Type":"ContainerDied","Data":"2b6870dfc6a6aa4776bad120ab2416e26e4267b1479e1694092999e8e19d0a5d"} Dec 01 21:53:49 crc kubenswrapper[4857]: I1201 21:53:49.508972 4857 generic.go:334] "Generic (PLEG): container finished" podID="55e4213d-69d8-4f02-adcb-e2249fb3dccf" containerID="2361a36edb8bf39369dbc352f577e586f8366a33a92ad0d587cced04acc51a77" exitCode=0 Dec 01 21:53:49 crc kubenswrapper[4857]: I1201 21:53:49.509057 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5zzh6" event={"ID":"55e4213d-69d8-4f02-adcb-e2249fb3dccf","Type":"ContainerDied","Data":"2361a36edb8bf39369dbc352f577e586f8366a33a92ad0d587cced04acc51a77"} Dec 01 21:53:50 crc kubenswrapper[4857]: I1201 21:53:50.997111 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5zzh6" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.001640 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-41ef-account-create-update-rfn85" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.081621 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt6dl\" (UniqueName: \"kubernetes.io/projected/55e4213d-69d8-4f02-adcb-e2249fb3dccf-kube-api-access-dt6dl\") pod \"55e4213d-69d8-4f02-adcb-e2249fb3dccf\" (UID: \"55e4213d-69d8-4f02-adcb-e2249fb3dccf\") " Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.081672 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55e4213d-69d8-4f02-adcb-e2249fb3dccf-operator-scripts\") pod \"55e4213d-69d8-4f02-adcb-e2249fb3dccf\" (UID: \"55e4213d-69d8-4f02-adcb-e2249fb3dccf\") " Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.081723 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47h69\" (UniqueName: \"kubernetes.io/projected/b6fbba64-3540-42d5-a1dd-c5e65d4e1330-kube-api-access-47h69\") pod \"b6fbba64-3540-42d5-a1dd-c5e65d4e1330\" (UID: \"b6fbba64-3540-42d5-a1dd-c5e65d4e1330\") " Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.081764 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6fbba64-3540-42d5-a1dd-c5e65d4e1330-operator-scripts\") pod \"b6fbba64-3540-42d5-a1dd-c5e65d4e1330\" (UID: \"b6fbba64-3540-42d5-a1dd-c5e65d4e1330\") " Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.082758 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6fbba64-3540-42d5-a1dd-c5e65d4e1330-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b6fbba64-3540-42d5-a1dd-c5e65d4e1330" (UID: "b6fbba64-3540-42d5-a1dd-c5e65d4e1330"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.083328 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55e4213d-69d8-4f02-adcb-e2249fb3dccf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "55e4213d-69d8-4f02-adcb-e2249fb3dccf" (UID: "55e4213d-69d8-4f02-adcb-e2249fb3dccf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.104149 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6fbba64-3540-42d5-a1dd-c5e65d4e1330-kube-api-access-47h69" (OuterVolumeSpecName: "kube-api-access-47h69") pod "b6fbba64-3540-42d5-a1dd-c5e65d4e1330" (UID: "b6fbba64-3540-42d5-a1dd-c5e65d4e1330"). InnerVolumeSpecName "kube-api-access-47h69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.104236 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55e4213d-69d8-4f02-adcb-e2249fb3dccf-kube-api-access-dt6dl" (OuterVolumeSpecName: "kube-api-access-dt6dl") pod "55e4213d-69d8-4f02-adcb-e2249fb3dccf" (UID: "55e4213d-69d8-4f02-adcb-e2249fb3dccf"). InnerVolumeSpecName "kube-api-access-dt6dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.183908 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55e4213d-69d8-4f02-adcb-e2249fb3dccf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.183977 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47h69\" (UniqueName: \"kubernetes.io/projected/b6fbba64-3540-42d5-a1dd-c5e65d4e1330-kube-api-access-47h69\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.184000 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6fbba64-3540-42d5-a1dd-c5e65d4e1330-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.184025 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt6dl\" (UniqueName: \"kubernetes.io/projected/55e4213d-69d8-4f02-adcb-e2249fb3dccf-kube-api-access-dt6dl\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.285610 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:53:51 crc kubenswrapper[4857]: E1201 21:53:51.285864 4857 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 21:53:51 crc kubenswrapper[4857]: E1201 21:53:51.285901 4857 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 21:53:51 crc kubenswrapper[4857]: E1201 21:53:51.285975 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift podName:3ea80def-31e0-4cc2-9f15-26ee35c62641 nodeName:}" failed. No retries permitted until 2025-12-01 21:54:07.285951371 +0000 UTC m=+1165.776013698 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift") pod "swift-storage-0" (UID: "3ea80def-31e0-4cc2-9f15-26ee35c62641") : configmap "swift-ring-files" not found Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.528393 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5zzh6" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.528574 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5zzh6" event={"ID":"55e4213d-69d8-4f02-adcb-e2249fb3dccf","Type":"ContainerDied","Data":"54718abecf0dcd430b9af7ca623a8913d80dd1406d615f6e4f6b9a869fc1e0af"} Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.529315 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54718abecf0dcd430b9af7ca623a8913d80dd1406d615f6e4f6b9a869fc1e0af" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.530721 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-41ef-account-create-update-rfn85" event={"ID":"b6fbba64-3540-42d5-a1dd-c5e65d4e1330","Type":"ContainerDied","Data":"319f4374adefbf6af796b214a2d291c12a3496571d0ce618f667f73d40c6276d"} Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.530764 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="319f4374adefbf6af796b214a2d291c12a3496571d0ce618f667f73d40c6276d" Dec 01 21:53:51 crc kubenswrapper[4857]: I1201 21:53:51.530807 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-41ef-account-create-update-rfn85" Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.470645 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nhfhn" podUID="32213ad3-4287-4968-9d42-bfeb2ff558d1" containerName="ovn-controller" probeResult="failure" output=< Dec 01 21:53:52 crc kubenswrapper[4857]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 21:53:52 crc kubenswrapper[4857]: > Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.598397 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.602820 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gwbmr" Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.854741 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-nhfhn-config-27vds"] Dec 01 21:53:52 crc kubenswrapper[4857]: E1201 21:53:52.855567 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e4213d-69d8-4f02-adcb-e2249fb3dccf" containerName="mariadb-database-create" Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.855671 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e4213d-69d8-4f02-adcb-e2249fb3dccf" containerName="mariadb-database-create" Dec 01 21:53:52 crc kubenswrapper[4857]: E1201 21:53:52.855757 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6fbba64-3540-42d5-a1dd-c5e65d4e1330" containerName="mariadb-account-create-update" Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.855826 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6fbba64-3540-42d5-a1dd-c5e65d4e1330" containerName="mariadb-account-create-update" Dec 01 21:53:52 crc kubenswrapper[4857]: E1201 21:53:52.855923 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31eea916-82d4-4918-8577-68276ba53766" containerName="mariadb-account-create-update" Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.855998 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="31eea916-82d4-4918-8577-68276ba53766" containerName="mariadb-account-create-update" Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.856289 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="31eea916-82d4-4918-8577-68276ba53766" containerName="mariadb-account-create-update" Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.856410 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6fbba64-3540-42d5-a1dd-c5e65d4e1330" containerName="mariadb-account-create-update" Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.856500 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="55e4213d-69d8-4f02-adcb-e2249fb3dccf" containerName="mariadb-database-create" Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.857207 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.860310 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 01 21:53:52 crc kubenswrapper[4857]: I1201 21:53:52.872108 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nhfhn-config-27vds"] Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.010941 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/71c72894-5bf7-4a1a-8aac-b15708c534e0-additional-scripts\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.011007 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-run-ovn\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.011077 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71c72894-5bf7-4a1a-8aac-b15708c534e0-scripts\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.011120 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v84r\" (UniqueName: \"kubernetes.io/projected/71c72894-5bf7-4a1a-8aac-b15708c534e0-kube-api-access-9v84r\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.011221 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-run\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.011258 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-log-ovn\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.112388 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71c72894-5bf7-4a1a-8aac-b15708c534e0-scripts\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.112786 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v84r\" (UniqueName: \"kubernetes.io/projected/71c72894-5bf7-4a1a-8aac-b15708c534e0-kube-api-access-9v84r\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.112855 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-run\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.112896 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-log-ovn\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.113010 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/71c72894-5bf7-4a1a-8aac-b15708c534e0-additional-scripts\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.113106 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-run-ovn\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.113213 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-run\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.113289 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-run-ovn\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.113452 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-log-ovn\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.113890 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/71c72894-5bf7-4a1a-8aac-b15708c534e0-additional-scripts\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.114955 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71c72894-5bf7-4a1a-8aac-b15708c534e0-scripts\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.136374 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v84r\" (UniqueName: \"kubernetes.io/projected/71c72894-5bf7-4a1a-8aac-b15708c534e0-kube-api-access-9v84r\") pod \"ovn-controller-nhfhn-config-27vds\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.184584 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.556271 4857 generic.go:334] "Generic (PLEG): container finished" podID="34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b" containerID="476ec1f85af117a366e28b21da443ea58ff745bcf0964eddef9c6cdf44627d43" exitCode=0 Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.556309 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-q5szn" event={"ID":"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b","Type":"ContainerDied","Data":"476ec1f85af117a366e28b21da443ea58ff745bcf0964eddef9c6cdf44627d43"} Dec 01 21:53:53 crc kubenswrapper[4857]: I1201 21:53:53.768683 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nhfhn-config-27vds"] Dec 01 21:53:53 crc kubenswrapper[4857]: W1201 21:53:53.773782 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71c72894_5bf7_4a1a_8aac_b15708c534e0.slice/crio-3715695b8029612e4bb748ecd66c09cdf8676b584d6f4934f3b151d20687f835 WatchSource:0}: Error finding container 3715695b8029612e4bb748ecd66c09cdf8676b584d6f4934f3b151d20687f835: Status 404 returned error can't find the container with id 3715695b8029612e4bb748ecd66c09cdf8676b584d6f4934f3b151d20687f835 Dec 01 21:53:54 crc kubenswrapper[4857]: I1201 21:53:54.580590 4857 generic.go:334] "Generic (PLEG): container finished" podID="71c72894-5bf7-4a1a-8aac-b15708c534e0" containerID="08785a340aa8e4a8be90a3d1f8822a6afbed843f3a156732aa378592d9c0061f" exitCode=0 Dec 01 21:53:54 crc kubenswrapper[4857]: I1201 21:53:54.581655 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nhfhn-config-27vds" event={"ID":"71c72894-5bf7-4a1a-8aac-b15708c534e0","Type":"ContainerDied","Data":"08785a340aa8e4a8be90a3d1f8822a6afbed843f3a156732aa378592d9c0061f"} Dec 01 21:53:54 crc kubenswrapper[4857]: I1201 21:53:54.581687 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nhfhn-config-27vds" event={"ID":"71c72894-5bf7-4a1a-8aac-b15708c534e0","Type":"ContainerStarted","Data":"3715695b8029612e4bb748ecd66c09cdf8676b584d6f4934f3b151d20687f835"} Dec 01 21:53:54 crc kubenswrapper[4857]: I1201 21:53:54.929203 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.052779 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-combined-ca-bundle\") pod \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.052861 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-swiftconf\") pod \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.052915 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-scripts\") pod \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.053001 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-dispersionconf\") pod \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.053102 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-ring-data-devices\") pod \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.053182 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-etc-swift\") pod \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.053266 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhcln\" (UniqueName: \"kubernetes.io/projected/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-kube-api-access-rhcln\") pod \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\" (UID: \"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b\") " Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.055084 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b" (UID: "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.055446 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b" (UID: "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.067828 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-kube-api-access-rhcln" (OuterVolumeSpecName: "kube-api-access-rhcln") pod "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b" (UID: "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b"). InnerVolumeSpecName "kube-api-access-rhcln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.069000 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b" (UID: "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.088722 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-scripts" (OuterVolumeSpecName: "scripts") pod "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b" (UID: "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.100662 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b" (UID: "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.108931 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b" (UID: "34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.155554 4857 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.155596 4857 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.155609 4857 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.155620 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhcln\" (UniqueName: \"kubernetes.io/projected/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-kube-api-access-rhcln\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.155631 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.155640 4857 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.155651 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.596709 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-q5szn" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.596700 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-q5szn" event={"ID":"34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b","Type":"ContainerDied","Data":"325363d28ef0489f5d1239a9cd4c5c8a88677d7cd6769065c716943eb44cc751"} Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.596901 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="325363d28ef0489f5d1239a9cd4c5c8a88677d7cd6769065c716943eb44cc751" Dec 01 21:53:55 crc kubenswrapper[4857]: I1201 21:53:55.946203 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.072417 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71c72894-5bf7-4a1a-8aac-b15708c534e0-scripts\") pod \"71c72894-5bf7-4a1a-8aac-b15708c534e0\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.072509 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/71c72894-5bf7-4a1a-8aac-b15708c534e0-additional-scripts\") pod \"71c72894-5bf7-4a1a-8aac-b15708c534e0\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.072540 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v84r\" (UniqueName: \"kubernetes.io/projected/71c72894-5bf7-4a1a-8aac-b15708c534e0-kube-api-access-9v84r\") pod \"71c72894-5bf7-4a1a-8aac-b15708c534e0\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.072834 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-run" (OuterVolumeSpecName: "var-run") pod "71c72894-5bf7-4a1a-8aac-b15708c534e0" (UID: "71c72894-5bf7-4a1a-8aac-b15708c534e0"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.073646 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-run\") pod \"71c72894-5bf7-4a1a-8aac-b15708c534e0\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.073713 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71c72894-5bf7-4a1a-8aac-b15708c534e0-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "71c72894-5bf7-4a1a-8aac-b15708c534e0" (UID: "71c72894-5bf7-4a1a-8aac-b15708c534e0"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.073757 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "71c72894-5bf7-4a1a-8aac-b15708c534e0" (UID: "71c72894-5bf7-4a1a-8aac-b15708c534e0"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.073733 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-log-ovn\") pod \"71c72894-5bf7-4a1a-8aac-b15708c534e0\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.073862 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-run-ovn\") pod \"71c72894-5bf7-4a1a-8aac-b15708c534e0\" (UID: \"71c72894-5bf7-4a1a-8aac-b15708c534e0\") " Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.074104 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "71c72894-5bf7-4a1a-8aac-b15708c534e0" (UID: "71c72894-5bf7-4a1a-8aac-b15708c534e0"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.074314 4857 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/71c72894-5bf7-4a1a-8aac-b15708c534e0-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.074341 4857 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-run\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.074350 4857 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.074359 4857 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/71c72894-5bf7-4a1a-8aac-b15708c534e0-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.074301 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71c72894-5bf7-4a1a-8aac-b15708c534e0-scripts" (OuterVolumeSpecName: "scripts") pod "71c72894-5bf7-4a1a-8aac-b15708c534e0" (UID: "71c72894-5bf7-4a1a-8aac-b15708c534e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.135492 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71c72894-5bf7-4a1a-8aac-b15708c534e0-kube-api-access-9v84r" (OuterVolumeSpecName: "kube-api-access-9v84r") pod "71c72894-5bf7-4a1a-8aac-b15708c534e0" (UID: "71c72894-5bf7-4a1a-8aac-b15708c534e0"). InnerVolumeSpecName "kube-api-access-9v84r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.175857 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71c72894-5bf7-4a1a-8aac-b15708c534e0-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.175893 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v84r\" (UniqueName: \"kubernetes.io/projected/71c72894-5bf7-4a1a-8aac-b15708c534e0-kube-api-access-9v84r\") on node \"crc\" DevicePath \"\"" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.608858 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nhfhn-config-27vds" event={"ID":"71c72894-5bf7-4a1a-8aac-b15708c534e0","Type":"ContainerDied","Data":"3715695b8029612e4bb748ecd66c09cdf8676b584d6f4934f3b151d20687f835"} Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.608917 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nhfhn-config-27vds" Dec 01 21:53:56 crc kubenswrapper[4857]: I1201 21:53:56.608926 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3715695b8029612e4bb748ecd66c09cdf8676b584d6f4934f3b151d20687f835" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.086017 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-nhfhn-config-27vds"] Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.097466 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-nhfhn-config-27vds"] Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.270830 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-9f69d"] Dec 01 21:53:57 crc kubenswrapper[4857]: E1201 21:53:57.271579 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71c72894-5bf7-4a1a-8aac-b15708c534e0" containerName="ovn-config" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.271593 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="71c72894-5bf7-4a1a-8aac-b15708c534e0" containerName="ovn-config" Dec 01 21:53:57 crc kubenswrapper[4857]: E1201 21:53:57.271618 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b" containerName="swift-ring-rebalance" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.271624 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b" containerName="swift-ring-rebalance" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.271805 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b" containerName="swift-ring-rebalance" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.271827 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="71c72894-5bf7-4a1a-8aac-b15708c534e0" containerName="ovn-config" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.272401 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.276628 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.276859 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bjgmp" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.287220 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-9f69d"] Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.358816 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6tbd\" (UniqueName: \"kubernetes.io/projected/f985748e-24cd-4d1b-abe8-7a579aee87af-kube-api-access-d6tbd\") pod \"glance-db-sync-9f69d\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.359000 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-config-data\") pod \"glance-db-sync-9f69d\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.359129 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-combined-ca-bundle\") pod \"glance-db-sync-9f69d\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.359174 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-db-sync-config-data\") pod \"glance-db-sync-9f69d\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.459141 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-nhfhn" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.460833 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-config-data\") pod \"glance-db-sync-9f69d\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.460952 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-combined-ca-bundle\") pod \"glance-db-sync-9f69d\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.460993 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-db-sync-config-data\") pod \"glance-db-sync-9f69d\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.461011 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6tbd\" (UniqueName: \"kubernetes.io/projected/f985748e-24cd-4d1b-abe8-7a579aee87af-kube-api-access-d6tbd\") pod \"glance-db-sync-9f69d\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.467761 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-config-data\") pod \"glance-db-sync-9f69d\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.473563 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-db-sync-config-data\") pod \"glance-db-sync-9f69d\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.481668 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6tbd\" (UniqueName: \"kubernetes.io/projected/f985748e-24cd-4d1b-abe8-7a579aee87af-kube-api-access-d6tbd\") pod \"glance-db-sync-9f69d\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.483439 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-combined-ca-bundle\") pod \"glance-db-sync-9f69d\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.605704 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9f69d" Dec 01 21:53:57 crc kubenswrapper[4857]: I1201 21:53:57.848137 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71c72894-5bf7-4a1a-8aac-b15708c534e0" path="/var/lib/kubelet/pods/71c72894-5bf7-4a1a-8aac-b15708c534e0/volumes" Dec 01 21:53:58 crc kubenswrapper[4857]: I1201 21:53:58.015305 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-9f69d"] Dec 01 21:53:58 crc kubenswrapper[4857]: I1201 21:53:58.628804 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9f69d" event={"ID":"f985748e-24cd-4d1b-abe8-7a579aee87af","Type":"ContainerStarted","Data":"9b28193c25aa92a56c9a52ba9527e67c2240be38285a977841a136cf67843917"} Dec 01 21:53:58 crc kubenswrapper[4857]: I1201 21:53:58.775890 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="235e0415-1e77-480d-b3da-157ed0252d4f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 01 21:53:59 crc kubenswrapper[4857]: I1201 21:53:59.708187 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="84a09f7d-b0a7-4055-93e1-83d11edb6467" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 01 21:54:07 crc kubenswrapper[4857]: I1201 21:54:07.354657 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:54:07 crc kubenswrapper[4857]: I1201 21:54:07.372911 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ea80def-31e0-4cc2-9f15-26ee35c62641-etc-swift\") pod \"swift-storage-0\" (UID: \"3ea80def-31e0-4cc2-9f15-26ee35c62641\") " pod="openstack/swift-storage-0" Dec 01 21:54:07 crc kubenswrapper[4857]: I1201 21:54:07.534637 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 01 21:54:08 crc kubenswrapper[4857]: I1201 21:54:08.776559 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.156081 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-qm6hg"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.157551 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-qm6hg" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.166662 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-qm6hg"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.241256 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-v2h7r"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.243354 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v2h7r" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.253176 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-v2h7r"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.290901 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/018d4544-4c86-45b0-a320-77d857f6e49b-operator-scripts\") pod \"heat-db-create-qm6hg\" (UID: \"018d4544-4c86-45b0-a320-77d857f6e49b\") " pod="openstack/heat-db-create-qm6hg" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.291260 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvt4t\" (UniqueName: \"kubernetes.io/projected/a6d47cc8-65ed-4b57-abb2-2216c9ed451a-kube-api-access-mvt4t\") pod \"cinder-db-create-v2h7r\" (UID: \"a6d47cc8-65ed-4b57-abb2-2216c9ed451a\") " pod="openstack/cinder-db-create-v2h7r" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.291312 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhlg7\" (UniqueName: \"kubernetes.io/projected/018d4544-4c86-45b0-a320-77d857f6e49b-kube-api-access-xhlg7\") pod \"heat-db-create-qm6hg\" (UID: \"018d4544-4c86-45b0-a320-77d857f6e49b\") " pod="openstack/heat-db-create-qm6hg" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.291609 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6d47cc8-65ed-4b57-abb2-2216c9ed451a-operator-scripts\") pod \"cinder-db-create-v2h7r\" (UID: \"a6d47cc8-65ed-4b57-abb2-2216c9ed451a\") " pod="openstack/cinder-db-create-v2h7r" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.358219 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-27ab-account-create-update-4mr2x"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.360747 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-27ab-account-create-update-4mr2x" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.362979 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.370695 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-27ab-account-create-update-4mr2x"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.396012 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvt4t\" (UniqueName: \"kubernetes.io/projected/a6d47cc8-65ed-4b57-abb2-2216c9ed451a-kube-api-access-mvt4t\") pod \"cinder-db-create-v2h7r\" (UID: \"a6d47cc8-65ed-4b57-abb2-2216c9ed451a\") " pod="openstack/cinder-db-create-v2h7r" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.396068 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhlg7\" (UniqueName: \"kubernetes.io/projected/018d4544-4c86-45b0-a320-77d857f6e49b-kube-api-access-xhlg7\") pod \"heat-db-create-qm6hg\" (UID: \"018d4544-4c86-45b0-a320-77d857f6e49b\") " pod="openstack/heat-db-create-qm6hg" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.396139 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6d47cc8-65ed-4b57-abb2-2216c9ed451a-operator-scripts\") pod \"cinder-db-create-v2h7r\" (UID: \"a6d47cc8-65ed-4b57-abb2-2216c9ed451a\") " pod="openstack/cinder-db-create-v2h7r" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.396174 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/018d4544-4c86-45b0-a320-77d857f6e49b-operator-scripts\") pod \"heat-db-create-qm6hg\" (UID: \"018d4544-4c86-45b0-a320-77d857f6e49b\") " pod="openstack/heat-db-create-qm6hg" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.397224 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/018d4544-4c86-45b0-a320-77d857f6e49b-operator-scripts\") pod \"heat-db-create-qm6hg\" (UID: \"018d4544-4c86-45b0-a320-77d857f6e49b\") " pod="openstack/heat-db-create-qm6hg" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.398344 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6d47cc8-65ed-4b57-abb2-2216c9ed451a-operator-scripts\") pod \"cinder-db-create-v2h7r\" (UID: \"a6d47cc8-65ed-4b57-abb2-2216c9ed451a\") " pod="openstack/cinder-db-create-v2h7r" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.517597 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvt4t\" (UniqueName: \"kubernetes.io/projected/a6d47cc8-65ed-4b57-abb2-2216c9ed451a-kube-api-access-mvt4t\") pod \"cinder-db-create-v2h7r\" (UID: \"a6d47cc8-65ed-4b57-abb2-2216c9ed451a\") " pod="openstack/cinder-db-create-v2h7r" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.518400 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf8st\" (UniqueName: \"kubernetes.io/projected/87ab7c0d-04a8-4527-966d-d48abfc3c645-kube-api-access-vf8st\") pod \"cinder-27ab-account-create-update-4mr2x\" (UID: \"87ab7c0d-04a8-4527-966d-d48abfc3c645\") " pod="openstack/cinder-27ab-account-create-update-4mr2x" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.518551 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87ab7c0d-04a8-4527-966d-d48abfc3c645-operator-scripts\") pod \"cinder-27ab-account-create-update-4mr2x\" (UID: \"87ab7c0d-04a8-4527-966d-d48abfc3c645\") " pod="openstack/cinder-27ab-account-create-update-4mr2x" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.592951 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v2h7r" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.593310 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhlg7\" (UniqueName: \"kubernetes.io/projected/018d4544-4c86-45b0-a320-77d857f6e49b-kube-api-access-xhlg7\") pod \"heat-db-create-qm6hg\" (UID: \"018d4544-4c86-45b0-a320-77d857f6e49b\") " pod="openstack/heat-db-create-qm6hg" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.599390 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-zdjss"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.600611 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zdjss" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.608465 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-2fb1-account-create-update-66k42"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.609206 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2fb1-account-create-update-66k42" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.612881 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.615320 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-zdjss"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.619835 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf8st\" (UniqueName: \"kubernetes.io/projected/87ab7c0d-04a8-4527-966d-d48abfc3c645-kube-api-access-vf8st\") pod \"cinder-27ab-account-create-update-4mr2x\" (UID: \"87ab7c0d-04a8-4527-966d-d48abfc3c645\") " pod="openstack/cinder-27ab-account-create-update-4mr2x" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.619920 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87ab7c0d-04a8-4527-966d-d48abfc3c645-operator-scripts\") pod \"cinder-27ab-account-create-update-4mr2x\" (UID: \"87ab7c0d-04a8-4527-966d-d48abfc3c645\") " pod="openstack/cinder-27ab-account-create-update-4mr2x" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.620667 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87ab7c0d-04a8-4527-966d-d48abfc3c645-operator-scripts\") pod \"cinder-27ab-account-create-update-4mr2x\" (UID: \"87ab7c0d-04a8-4527-966d-d48abfc3c645\") " pod="openstack/cinder-27ab-account-create-update-4mr2x" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.628643 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2fb1-account-create-update-66k42"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.643200 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-a859-account-create-update-z4slz"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.645602 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a859-account-create-update-z4slz" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.656830 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-a859-account-create-update-z4slz"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.657350 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.658930 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf8st\" (UniqueName: \"kubernetes.io/projected/87ab7c0d-04a8-4527-966d-d48abfc3c645-kube-api-access-vf8st\") pod \"cinder-27ab-account-create-update-4mr2x\" (UID: \"87ab7c0d-04a8-4527-966d-d48abfc3c645\") " pod="openstack/cinder-27ab-account-create-update-4mr2x" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.659001 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-9gzqb"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.660180 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.666364 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.666469 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6ll6v" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.667145 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.667282 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.680157 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-27ab-account-create-update-4mr2x" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.681701 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-9gzqb"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.686270 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.722853 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1c0aebf-0263-4947-88ee-70a119ef3453-operator-scripts\") pod \"barbican-2fb1-account-create-update-66k42\" (UID: \"a1c0aebf-0263-4947-88ee-70a119ef3453\") " pod="openstack/barbican-2fb1-account-create-update-66k42" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.722931 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-combined-ca-bundle\") pod \"keystone-db-sync-9gzqb\" (UID: \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\") " pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.722956 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkmqk\" (UniqueName: \"kubernetes.io/projected/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-kube-api-access-bkmqk\") pod \"keystone-db-sync-9gzqb\" (UID: \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\") " pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.723090 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2sk9\" (UniqueName: \"kubernetes.io/projected/8107394e-a9a2-43a6-bf43-3c785c0455c6-kube-api-access-z2sk9\") pod \"heat-a859-account-create-update-z4slz\" (UID: \"8107394e-a9a2-43a6-bf43-3c785c0455c6\") " pod="openstack/heat-a859-account-create-update-z4slz" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.723409 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqdld\" (UniqueName: \"kubernetes.io/projected/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec-kube-api-access-dqdld\") pod \"barbican-db-create-zdjss\" (UID: \"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec\") " pod="openstack/barbican-db-create-zdjss" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.723481 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8107394e-a9a2-43a6-bf43-3c785c0455c6-operator-scripts\") pod \"heat-a859-account-create-update-z4slz\" (UID: \"8107394e-a9a2-43a6-bf43-3c785c0455c6\") " pod="openstack/heat-a859-account-create-update-z4slz" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.723506 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95khn\" (UniqueName: \"kubernetes.io/projected/a1c0aebf-0263-4947-88ee-70a119ef3453-kube-api-access-95khn\") pod \"barbican-2fb1-account-create-update-66k42\" (UID: \"a1c0aebf-0263-4947-88ee-70a119ef3453\") " pod="openstack/barbican-2fb1-account-create-update-66k42" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.723829 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec-operator-scripts\") pod \"barbican-db-create-zdjss\" (UID: \"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec\") " pod="openstack/barbican-db-create-zdjss" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.724141 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-config-data\") pod \"keystone-db-sync-9gzqb\" (UID: \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\") " pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.776622 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-qm6hg" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.826276 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqdld\" (UniqueName: \"kubernetes.io/projected/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec-kube-api-access-dqdld\") pod \"barbican-db-create-zdjss\" (UID: \"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec\") " pod="openstack/barbican-db-create-zdjss" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.826350 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8107394e-a9a2-43a6-bf43-3c785c0455c6-operator-scripts\") pod \"heat-a859-account-create-update-z4slz\" (UID: \"8107394e-a9a2-43a6-bf43-3c785c0455c6\") " pod="openstack/heat-a859-account-create-update-z4slz" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.826379 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95khn\" (UniqueName: \"kubernetes.io/projected/a1c0aebf-0263-4947-88ee-70a119ef3453-kube-api-access-95khn\") pod \"barbican-2fb1-account-create-update-66k42\" (UID: \"a1c0aebf-0263-4947-88ee-70a119ef3453\") " pod="openstack/barbican-2fb1-account-create-update-66k42" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.826420 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec-operator-scripts\") pod \"barbican-db-create-zdjss\" (UID: \"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec\") " pod="openstack/barbican-db-create-zdjss" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.826912 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-config-data\") pod \"keystone-db-sync-9gzqb\" (UID: \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\") " pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.827647 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8107394e-a9a2-43a6-bf43-3c785c0455c6-operator-scripts\") pod \"heat-a859-account-create-update-z4slz\" (UID: \"8107394e-a9a2-43a6-bf43-3c785c0455c6\") " pod="openstack/heat-a859-account-create-update-z4slz" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.827730 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec-operator-scripts\") pod \"barbican-db-create-zdjss\" (UID: \"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec\") " pod="openstack/barbican-db-create-zdjss" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.831234 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1c0aebf-0263-4947-88ee-70a119ef3453-operator-scripts\") pod \"barbican-2fb1-account-create-update-66k42\" (UID: \"a1c0aebf-0263-4947-88ee-70a119ef3453\") " pod="openstack/barbican-2fb1-account-create-update-66k42" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.831270 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-combined-ca-bundle\") pod \"keystone-db-sync-9gzqb\" (UID: \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\") " pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.831300 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkmqk\" (UniqueName: \"kubernetes.io/projected/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-kube-api-access-bkmqk\") pod \"keystone-db-sync-9gzqb\" (UID: \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\") " pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.832069 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1c0aebf-0263-4947-88ee-70a119ef3453-operator-scripts\") pod \"barbican-2fb1-account-create-update-66k42\" (UID: \"a1c0aebf-0263-4947-88ee-70a119ef3453\") " pod="openstack/barbican-2fb1-account-create-update-66k42" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.834187 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2sk9\" (UniqueName: \"kubernetes.io/projected/8107394e-a9a2-43a6-bf43-3c785c0455c6-kube-api-access-z2sk9\") pod \"heat-a859-account-create-update-z4slz\" (UID: \"8107394e-a9a2-43a6-bf43-3c785c0455c6\") " pod="openstack/heat-a859-account-create-update-z4slz" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.834943 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-combined-ca-bundle\") pod \"keystone-db-sync-9gzqb\" (UID: \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\") " pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.858109 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-config-data\") pod \"keystone-db-sync-9gzqb\" (UID: \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\") " pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.860175 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-9k4pr"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.862235 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9k4pr" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.862343 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqdld\" (UniqueName: \"kubernetes.io/projected/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec-kube-api-access-dqdld\") pod \"barbican-db-create-zdjss\" (UID: \"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec\") " pod="openstack/barbican-db-create-zdjss" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.868085 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95khn\" (UniqueName: \"kubernetes.io/projected/a1c0aebf-0263-4947-88ee-70a119ef3453-kube-api-access-95khn\") pod \"barbican-2fb1-account-create-update-66k42\" (UID: \"a1c0aebf-0263-4947-88ee-70a119ef3453\") " pod="openstack/barbican-2fb1-account-create-update-66k42" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.871467 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkmqk\" (UniqueName: \"kubernetes.io/projected/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-kube-api-access-bkmqk\") pod \"keystone-db-sync-9gzqb\" (UID: \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\") " pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.884702 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2sk9\" (UniqueName: \"kubernetes.io/projected/8107394e-a9a2-43a6-bf43-3c785c0455c6-kube-api-access-z2sk9\") pod \"heat-a859-account-create-update-z4slz\" (UID: \"8107394e-a9a2-43a6-bf43-3c785c0455c6\") " pod="openstack/heat-a859-account-create-update-z4slz" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.922159 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9k4pr"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.934478 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zdjss" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.938064 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85b66a20-d9b6-45ca-889b-87121ebac3f0-operator-scripts\") pod \"neutron-db-create-9k4pr\" (UID: \"85b66a20-d9b6-45ca-889b-87121ebac3f0\") " pod="openstack/neutron-db-create-9k4pr" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.938264 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t92h4\" (UniqueName: \"kubernetes.io/projected/85b66a20-d9b6-45ca-889b-87121ebac3f0-kube-api-access-t92h4\") pod \"neutron-db-create-9k4pr\" (UID: \"85b66a20-d9b6-45ca-889b-87121ebac3f0\") " pod="openstack/neutron-db-create-9k4pr" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.940523 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2fb1-account-create-update-66k42" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.951768 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-fac7-account-create-update-h6hzj"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.952939 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fac7-account-create-update-h6hzj" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.958015 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.988837 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-fac7-account-create-update-h6hzj"] Dec 01 21:54:09 crc kubenswrapper[4857]: I1201 21:54:09.991460 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a859-account-create-update-z4slz" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.000497 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.040457 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31e91520-3698-4c0a-82f7-9792c769f9f3-operator-scripts\") pod \"neutron-fac7-account-create-update-h6hzj\" (UID: \"31e91520-3698-4c0a-82f7-9792c769f9f3\") " pod="openstack/neutron-fac7-account-create-update-h6hzj" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.040533 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t92h4\" (UniqueName: \"kubernetes.io/projected/85b66a20-d9b6-45ca-889b-87121ebac3f0-kube-api-access-t92h4\") pod \"neutron-db-create-9k4pr\" (UID: \"85b66a20-d9b6-45ca-889b-87121ebac3f0\") " pod="openstack/neutron-db-create-9k4pr" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.040659 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwh26\" (UniqueName: \"kubernetes.io/projected/31e91520-3698-4c0a-82f7-9792c769f9f3-kube-api-access-lwh26\") pod \"neutron-fac7-account-create-update-h6hzj\" (UID: \"31e91520-3698-4c0a-82f7-9792c769f9f3\") " pod="openstack/neutron-fac7-account-create-update-h6hzj" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.040701 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85b66a20-d9b6-45ca-889b-87121ebac3f0-operator-scripts\") pod \"neutron-db-create-9k4pr\" (UID: \"85b66a20-d9b6-45ca-889b-87121ebac3f0\") " pod="openstack/neutron-db-create-9k4pr" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.041506 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85b66a20-d9b6-45ca-889b-87121ebac3f0-operator-scripts\") pod \"neutron-db-create-9k4pr\" (UID: \"85b66a20-d9b6-45ca-889b-87121ebac3f0\") " pod="openstack/neutron-db-create-9k4pr" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.058121 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t92h4\" (UniqueName: \"kubernetes.io/projected/85b66a20-d9b6-45ca-889b-87121ebac3f0-kube-api-access-t92h4\") pod \"neutron-db-create-9k4pr\" (UID: \"85b66a20-d9b6-45ca-889b-87121ebac3f0\") " pod="openstack/neutron-db-create-9k4pr" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.142161 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwh26\" (UniqueName: \"kubernetes.io/projected/31e91520-3698-4c0a-82f7-9792c769f9f3-kube-api-access-lwh26\") pod \"neutron-fac7-account-create-update-h6hzj\" (UID: \"31e91520-3698-4c0a-82f7-9792c769f9f3\") " pod="openstack/neutron-fac7-account-create-update-h6hzj" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.142366 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31e91520-3698-4c0a-82f7-9792c769f9f3-operator-scripts\") pod \"neutron-fac7-account-create-update-h6hzj\" (UID: \"31e91520-3698-4c0a-82f7-9792c769f9f3\") " pod="openstack/neutron-fac7-account-create-update-h6hzj" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.143387 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31e91520-3698-4c0a-82f7-9792c769f9f3-operator-scripts\") pod \"neutron-fac7-account-create-update-h6hzj\" (UID: \"31e91520-3698-4c0a-82f7-9792c769f9f3\") " pod="openstack/neutron-fac7-account-create-update-h6hzj" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.168077 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwh26\" (UniqueName: \"kubernetes.io/projected/31e91520-3698-4c0a-82f7-9792c769f9f3-kube-api-access-lwh26\") pod \"neutron-fac7-account-create-update-h6hzj\" (UID: \"31e91520-3698-4c0a-82f7-9792c769f9f3\") " pod="openstack/neutron-fac7-account-create-update-h6hzj" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.215874 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9k4pr" Dec 01 21:54:10 crc kubenswrapper[4857]: I1201 21:54:10.268733 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fac7-account-create-update-h6hzj" Dec 01 21:54:15 crc kubenswrapper[4857]: E1201 21:54:15.728410 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 01 21:54:15 crc kubenswrapper[4857]: E1201 21:54:15.729146 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d6tbd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-9f69d_openstack(f985748e-24cd-4d1b-abe8-7a579aee87af): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:54:15 crc kubenswrapper[4857]: E1201 21:54:15.739221 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-9f69d" podUID="f985748e-24cd-4d1b-abe8-7a579aee87af" Dec 01 21:54:15 crc kubenswrapper[4857]: E1201 21:54:15.850011 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-9f69d" podUID="f985748e-24cd-4d1b-abe8-7a579aee87af" Dec 01 21:54:16 crc kubenswrapper[4857]: I1201 21:54:16.742817 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-v2h7r"] Dec 01 21:54:16 crc kubenswrapper[4857]: I1201 21:54:16.856185 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-zdjss"] Dec 01 21:54:16 crc kubenswrapper[4857]: I1201 21:54:16.862007 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v2h7r" event={"ID":"a6d47cc8-65ed-4b57-abb2-2216c9ed451a","Type":"ContainerStarted","Data":"10c2d341c943a88e1133ba91799c762395fc3927b0b790f2798074e472b654b5"} Dec 01 21:54:16 crc kubenswrapper[4857]: I1201 21:54:16.864523 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9k4pr"] Dec 01 21:54:16 crc kubenswrapper[4857]: I1201 21:54:16.945297 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-fac7-account-create-update-h6hzj"] Dec 01 21:54:16 crc kubenswrapper[4857]: I1201 21:54:16.952201 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2fb1-account-create-update-66k42"] Dec 01 21:54:16 crc kubenswrapper[4857]: W1201 21:54:16.976369 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31e91520_3698_4c0a_82f7_9792c769f9f3.slice/crio-1a4d85672367c98e7747aa67d4256318e86c7960b573bb215a5e70ebd4adc36e WatchSource:0}: Error finding container 1a4d85672367c98e7747aa67d4256318e86c7960b573bb215a5e70ebd4adc36e: Status 404 returned error can't find the container with id 1a4d85672367c98e7747aa67d4256318e86c7960b573bb215a5e70ebd4adc36e Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.104291 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-9gzqb"] Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.120419 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-a859-account-create-update-z4slz"] Dec 01 21:54:17 crc kubenswrapper[4857]: W1201 21:54:17.124262 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddab7e48a_a3a1_4e9a_bfee_0e0fa712adab.slice/crio-cb9ac852a35faf061273d1f26f8c345b3288bad92344bfd76c134d27ef6e22ab WatchSource:0}: Error finding container cb9ac852a35faf061273d1f26f8c345b3288bad92344bfd76c134d27ef6e22ab: Status 404 returned error can't find the container with id cb9ac852a35faf061273d1f26f8c345b3288bad92344bfd76c134d27ef6e22ab Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.136138 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-27ab-account-create-update-4mr2x"] Dec 01 21:54:17 crc kubenswrapper[4857]: W1201 21:54:17.140195 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8107394e_a9a2_43a6_bf43_3c785c0455c6.slice/crio-fdc9638af2fdde2a55ffbdec4b269819feeae85ffd365b18c0a31e0f8e05bf8a WatchSource:0}: Error finding container fdc9638af2fdde2a55ffbdec4b269819feeae85ffd365b18c0a31e0f8e05bf8a: Status 404 returned error can't find the container with id fdc9638af2fdde2a55ffbdec4b269819feeae85ffd365b18c0a31e0f8e05bf8a Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.141837 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-qm6hg"] Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.217493 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 01 21:54:17 crc kubenswrapper[4857]: W1201 21:54:17.244731 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ea80def_31e0_4cc2_9f15_26ee35c62641.slice/crio-e9e55de7af2ef87c3bd4aceb16552e4075c2143e13051e8035f726ab359ad009 WatchSource:0}: Error finding container e9e55de7af2ef87c3bd4aceb16552e4075c2143e13051e8035f726ab359ad009: Status 404 returned error can't find the container with id e9e55de7af2ef87c3bd4aceb16552e4075c2143e13051e8035f726ab359ad009 Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.887446 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"e9e55de7af2ef87c3bd4aceb16552e4075c2143e13051e8035f726ab359ad009"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.890650 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-27ab-account-create-update-4mr2x" event={"ID":"87ab7c0d-04a8-4527-966d-d48abfc3c645","Type":"ContainerStarted","Data":"bcd3c4cc1fa08e49a17ac0058293eb0a8d38f872be03850ee36db36448ffc2e5"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.890673 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-27ab-account-create-update-4mr2x" event={"ID":"87ab7c0d-04a8-4527-966d-d48abfc3c645","Type":"ContainerStarted","Data":"490a70cddf8b63974fd1da14d70a1d6118f8b91e0160761b45b23f2307c4a696"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.893670 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a859-account-create-update-z4slz" event={"ID":"8107394e-a9a2-43a6-bf43-3c785c0455c6","Type":"ContainerStarted","Data":"bde0f719f33acb5825d68ec72a778784837af0491876344f6ef13c21fdcce410"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.893691 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a859-account-create-update-z4slz" event={"ID":"8107394e-a9a2-43a6-bf43-3c785c0455c6","Type":"ContainerStarted","Data":"fdc9638af2fdde2a55ffbdec4b269819feeae85ffd365b18c0a31e0f8e05bf8a"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.898332 4857 generic.go:334] "Generic (PLEG): container finished" podID="018d4544-4c86-45b0-a320-77d857f6e49b" containerID="3d7cc0a93806a18ef5ae6046437191792bd0c4c3a7bbf3d271e09fe3a296ab21" exitCode=0 Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.898427 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-qm6hg" event={"ID":"018d4544-4c86-45b0-a320-77d857f6e49b","Type":"ContainerDied","Data":"3d7cc0a93806a18ef5ae6046437191792bd0c4c3a7bbf3d271e09fe3a296ab21"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.898442 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-qm6hg" event={"ID":"018d4544-4c86-45b0-a320-77d857f6e49b","Type":"ContainerStarted","Data":"bde6500141f3d57db7ef2e6600785cacf0cde355d2d1a0e723b2f9a93f968211"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.900687 4857 generic.go:334] "Generic (PLEG): container finished" podID="5c8567c2-527e-4cba-9d7b-e3da31d2e0ec" containerID="84f9de00d10ea8061abd6045a4921193ab96a535804ec1b7ad5ad735a288be2b" exitCode=0 Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.900734 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-zdjss" event={"ID":"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec","Type":"ContainerDied","Data":"84f9de00d10ea8061abd6045a4921193ab96a535804ec1b7ad5ad735a288be2b"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.900945 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-zdjss" event={"ID":"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec","Type":"ContainerStarted","Data":"e5b6c2f04bccf71601458b37914f95492f4f0258bcadb1814402127125b4e9ff"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.902671 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2fb1-account-create-update-66k42" event={"ID":"a1c0aebf-0263-4947-88ee-70a119ef3453","Type":"ContainerStarted","Data":"b0697fa87148ec15461681265d17e3c82144f730492dfcb10678b693aac6413e"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.902691 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2fb1-account-create-update-66k42" event={"ID":"a1c0aebf-0263-4947-88ee-70a119ef3453","Type":"ContainerStarted","Data":"ee9f9cc0b9d307d884ffe96029f4922e3c723221b8c9e59500bb6d631d623102"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.905793 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fac7-account-create-update-h6hzj" event={"ID":"31e91520-3698-4c0a-82f7-9792c769f9f3","Type":"ContainerStarted","Data":"a48b2165996854d387a711a39f4527ae5f5798c6bc0cf92c2d60d20cb4d3a940"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.905827 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fac7-account-create-update-h6hzj" event={"ID":"31e91520-3698-4c0a-82f7-9792c769f9f3","Type":"ContainerStarted","Data":"1a4d85672367c98e7747aa67d4256318e86c7960b573bb215a5e70ebd4adc36e"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.913357 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v2h7r" event={"ID":"a6d47cc8-65ed-4b57-abb2-2216c9ed451a","Type":"ContainerStarted","Data":"eb1ba56055913886f8724efae2010400bc17d5261fd0374e49c8ae921124535c"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.915352 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9gzqb" event={"ID":"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab","Type":"ContainerStarted","Data":"cb9ac852a35faf061273d1f26f8c345b3288bad92344bfd76c134d27ef6e22ab"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.916186 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-27ab-account-create-update-4mr2x" podStartSLOduration=8.916173011 podStartE2EDuration="8.916173011s" podCreationTimestamp="2025-12-01 21:54:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:54:17.912376089 +0000 UTC m=+1176.402438416" watchObservedRunningTime="2025-12-01 21:54:17.916173011 +0000 UTC m=+1176.406235328" Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.917306 4857 generic.go:334] "Generic (PLEG): container finished" podID="85b66a20-d9b6-45ca-889b-87121ebac3f0" containerID="df651ca2cdaef9cd85a3e4cbe423f1711e4070912bb6df2273c3182a1c58435a" exitCode=0 Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.917337 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9k4pr" event={"ID":"85b66a20-d9b6-45ca-889b-87121ebac3f0","Type":"ContainerDied","Data":"df651ca2cdaef9cd85a3e4cbe423f1711e4070912bb6df2273c3182a1c58435a"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.917352 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9k4pr" event={"ID":"85b66a20-d9b6-45ca-889b-87121ebac3f0","Type":"ContainerStarted","Data":"8ac834de14b503b4edf50e16930f0e71e409ebb86a3a28a81290fa3870a11345"} Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.950549 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-fac7-account-create-update-h6hzj" podStartSLOduration=8.950524843 podStartE2EDuration="8.950524843s" podCreationTimestamp="2025-12-01 21:54:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:54:17.943255947 +0000 UTC m=+1176.433318274" watchObservedRunningTime="2025-12-01 21:54:17.950524843 +0000 UTC m=+1176.440587160" Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.984783 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-a859-account-create-update-z4slz" podStartSLOduration=8.984763072 podStartE2EDuration="8.984763072s" podCreationTimestamp="2025-12-01 21:54:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:54:17.978006708 +0000 UTC m=+1176.468069045" watchObservedRunningTime="2025-12-01 21:54:17.984763072 +0000 UTC m=+1176.474825389" Dec 01 21:54:17 crc kubenswrapper[4857]: I1201 21:54:17.999331 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-2fb1-account-create-update-66k42" podStartSLOduration=8.999318574 podStartE2EDuration="8.999318574s" podCreationTimestamp="2025-12-01 21:54:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:54:17.994576009 +0000 UTC m=+1176.484638326" watchObservedRunningTime="2025-12-01 21:54:17.999318574 +0000 UTC m=+1176.489380891" Dec 01 21:54:18 crc kubenswrapper[4857]: I1201 21:54:18.928835 4857 generic.go:334] "Generic (PLEG): container finished" podID="a6d47cc8-65ed-4b57-abb2-2216c9ed451a" containerID="eb1ba56055913886f8724efae2010400bc17d5261fd0374e49c8ae921124535c" exitCode=0 Dec 01 21:54:18 crc kubenswrapper[4857]: I1201 21:54:18.929308 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v2h7r" event={"ID":"a6d47cc8-65ed-4b57-abb2-2216c9ed451a","Type":"ContainerDied","Data":"eb1ba56055913886f8724efae2010400bc17d5261fd0374e49c8ae921124535c"} Dec 01 21:54:18 crc kubenswrapper[4857]: I1201 21:54:18.935623 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"d3c75288c05633edb5682eda8a6d4c39047ba2b5989605999a5bd399b76a47b3"} Dec 01 21:54:18 crc kubenswrapper[4857]: I1201 21:54:18.938237 4857 generic.go:334] "Generic (PLEG): container finished" podID="87ab7c0d-04a8-4527-966d-d48abfc3c645" containerID="bcd3c4cc1fa08e49a17ac0058293eb0a8d38f872be03850ee36db36448ffc2e5" exitCode=0 Dec 01 21:54:18 crc kubenswrapper[4857]: I1201 21:54:18.938372 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-27ab-account-create-update-4mr2x" event={"ID":"87ab7c0d-04a8-4527-966d-d48abfc3c645","Type":"ContainerDied","Data":"bcd3c4cc1fa08e49a17ac0058293eb0a8d38f872be03850ee36db36448ffc2e5"} Dec 01 21:54:18 crc kubenswrapper[4857]: I1201 21:54:18.941310 4857 generic.go:334] "Generic (PLEG): container finished" podID="8107394e-a9a2-43a6-bf43-3c785c0455c6" containerID="bde0f719f33acb5825d68ec72a778784837af0491876344f6ef13c21fdcce410" exitCode=0 Dec 01 21:54:18 crc kubenswrapper[4857]: I1201 21:54:18.941355 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a859-account-create-update-z4slz" event={"ID":"8107394e-a9a2-43a6-bf43-3c785c0455c6","Type":"ContainerDied","Data":"bde0f719f33acb5825d68ec72a778784837af0491876344f6ef13c21fdcce410"} Dec 01 21:54:18 crc kubenswrapper[4857]: I1201 21:54:18.943624 4857 generic.go:334] "Generic (PLEG): container finished" podID="a1c0aebf-0263-4947-88ee-70a119ef3453" containerID="b0697fa87148ec15461681265d17e3c82144f730492dfcb10678b693aac6413e" exitCode=0 Dec 01 21:54:18 crc kubenswrapper[4857]: I1201 21:54:18.943671 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2fb1-account-create-update-66k42" event={"ID":"a1c0aebf-0263-4947-88ee-70a119ef3453","Type":"ContainerDied","Data":"b0697fa87148ec15461681265d17e3c82144f730492dfcb10678b693aac6413e"} Dec 01 21:54:18 crc kubenswrapper[4857]: I1201 21:54:18.945423 4857 generic.go:334] "Generic (PLEG): container finished" podID="31e91520-3698-4c0a-82f7-9792c769f9f3" containerID="a48b2165996854d387a711a39f4527ae5f5798c6bc0cf92c2d60d20cb4d3a940" exitCode=0 Dec 01 21:54:18 crc kubenswrapper[4857]: I1201 21:54:18.945617 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fac7-account-create-update-h6hzj" event={"ID":"31e91520-3698-4c0a-82f7-9792c769f9f3","Type":"ContainerDied","Data":"a48b2165996854d387a711a39f4527ae5f5798c6bc0cf92c2d60d20cb4d3a940"} Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.977709 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2fb1-account-create-update-66k42" event={"ID":"a1c0aebf-0263-4947-88ee-70a119ef3453","Type":"ContainerDied","Data":"ee9f9cc0b9d307d884ffe96029f4922e3c723221b8c9e59500bb6d631d623102"} Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.978488 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee9f9cc0b9d307d884ffe96029f4922e3c723221b8c9e59500bb6d631d623102" Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.979599 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fac7-account-create-update-h6hzj" event={"ID":"31e91520-3698-4c0a-82f7-9792c769f9f3","Type":"ContainerDied","Data":"1a4d85672367c98e7747aa67d4256318e86c7960b573bb215a5e70ebd4adc36e"} Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.979627 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a4d85672367c98e7747aa67d4256318e86c7960b573bb215a5e70ebd4adc36e" Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.980661 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v2h7r" event={"ID":"a6d47cc8-65ed-4b57-abb2-2216c9ed451a","Type":"ContainerDied","Data":"10c2d341c943a88e1133ba91799c762395fc3927b0b790f2798074e472b654b5"} Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.980699 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10c2d341c943a88e1133ba91799c762395fc3927b0b790f2798074e472b654b5" Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.982376 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9k4pr" event={"ID":"85b66a20-d9b6-45ca-889b-87121ebac3f0","Type":"ContainerDied","Data":"8ac834de14b503b4edf50e16930f0e71e409ebb86a3a28a81290fa3870a11345"} Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.982421 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ac834de14b503b4edf50e16930f0e71e409ebb86a3a28a81290fa3870a11345" Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.984029 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-qm6hg" event={"ID":"018d4544-4c86-45b0-a320-77d857f6e49b","Type":"ContainerDied","Data":"bde6500141f3d57db7ef2e6600785cacf0cde355d2d1a0e723b2f9a93f968211"} Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.984081 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bde6500141f3d57db7ef2e6600785cacf0cde355d2d1a0e723b2f9a93f968211" Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.985590 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-27ab-account-create-update-4mr2x" event={"ID":"87ab7c0d-04a8-4527-966d-d48abfc3c645","Type":"ContainerDied","Data":"490a70cddf8b63974fd1da14d70a1d6118f8b91e0160761b45b23f2307c4a696"} Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.985642 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="490a70cddf8b63974fd1da14d70a1d6118f8b91e0160761b45b23f2307c4a696" Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.987164 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-zdjss" event={"ID":"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec","Type":"ContainerDied","Data":"e5b6c2f04bccf71601458b37914f95492f4f0258bcadb1814402127125b4e9ff"} Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.987188 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5b6c2f04bccf71601458b37914f95492f4f0258bcadb1814402127125b4e9ff" Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.989443 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a859-account-create-update-z4slz" event={"ID":"8107394e-a9a2-43a6-bf43-3c785c0455c6","Type":"ContainerDied","Data":"fdc9638af2fdde2a55ffbdec4b269819feeae85ffd365b18c0a31e0f8e05bf8a"} Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.989463 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdc9638af2fdde2a55ffbdec4b269819feeae85ffd365b18c0a31e0f8e05bf8a" Dec 01 21:54:21 crc kubenswrapper[4857]: I1201 21:54:21.996852 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v2h7r" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.026082 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9k4pr" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.049836 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-qm6hg" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.080664 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2fb1-account-create-update-66k42" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.089351 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6d47cc8-65ed-4b57-abb2-2216c9ed451a-operator-scripts\") pod \"a6d47cc8-65ed-4b57-abb2-2216c9ed451a\" (UID: \"a6d47cc8-65ed-4b57-abb2-2216c9ed451a\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.089406 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvt4t\" (UniqueName: \"kubernetes.io/projected/a6d47cc8-65ed-4b57-abb2-2216c9ed451a-kube-api-access-mvt4t\") pod \"a6d47cc8-65ed-4b57-abb2-2216c9ed451a\" (UID: \"a6d47cc8-65ed-4b57-abb2-2216c9ed451a\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.090122 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6d47cc8-65ed-4b57-abb2-2216c9ed451a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a6d47cc8-65ed-4b57-abb2-2216c9ed451a" (UID: "a6d47cc8-65ed-4b57-abb2-2216c9ed451a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.100455 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a859-account-create-update-z4slz" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.107251 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6d47cc8-65ed-4b57-abb2-2216c9ed451a-kube-api-access-mvt4t" (OuterVolumeSpecName: "kube-api-access-mvt4t") pod "a6d47cc8-65ed-4b57-abb2-2216c9ed451a" (UID: "a6d47cc8-65ed-4b57-abb2-2216c9ed451a"). InnerVolumeSpecName "kube-api-access-mvt4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.159875 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zdjss" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.175282 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-27ab-account-create-update-4mr2x" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.178266 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fac7-account-create-update-h6hzj" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.190698 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t92h4\" (UniqueName: \"kubernetes.io/projected/85b66a20-d9b6-45ca-889b-87121ebac3f0-kube-api-access-t92h4\") pod \"85b66a20-d9b6-45ca-889b-87121ebac3f0\" (UID: \"85b66a20-d9b6-45ca-889b-87121ebac3f0\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.191012 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhlg7\" (UniqueName: \"kubernetes.io/projected/018d4544-4c86-45b0-a320-77d857f6e49b-kube-api-access-xhlg7\") pod \"018d4544-4c86-45b0-a320-77d857f6e49b\" (UID: \"018d4544-4c86-45b0-a320-77d857f6e49b\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.191106 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/018d4544-4c86-45b0-a320-77d857f6e49b-operator-scripts\") pod \"018d4544-4c86-45b0-a320-77d857f6e49b\" (UID: \"018d4544-4c86-45b0-a320-77d857f6e49b\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.191166 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1c0aebf-0263-4947-88ee-70a119ef3453-operator-scripts\") pod \"a1c0aebf-0263-4947-88ee-70a119ef3453\" (UID: \"a1c0aebf-0263-4947-88ee-70a119ef3453\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.191222 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2sk9\" (UniqueName: \"kubernetes.io/projected/8107394e-a9a2-43a6-bf43-3c785c0455c6-kube-api-access-z2sk9\") pod \"8107394e-a9a2-43a6-bf43-3c785c0455c6\" (UID: \"8107394e-a9a2-43a6-bf43-3c785c0455c6\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.191606 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85b66a20-d9b6-45ca-889b-87121ebac3f0-operator-scripts\") pod \"85b66a20-d9b6-45ca-889b-87121ebac3f0\" (UID: \"85b66a20-d9b6-45ca-889b-87121ebac3f0\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.191857 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1c0aebf-0263-4947-88ee-70a119ef3453-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a1c0aebf-0263-4947-88ee-70a119ef3453" (UID: "a1c0aebf-0263-4947-88ee-70a119ef3453"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.192534 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85b66a20-d9b6-45ca-889b-87121ebac3f0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85b66a20-d9b6-45ca-889b-87121ebac3f0" (UID: "85b66a20-d9b6-45ca-889b-87121ebac3f0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.192684 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8107394e-a9a2-43a6-bf43-3c785c0455c6-operator-scripts\") pod \"8107394e-a9a2-43a6-bf43-3c785c0455c6\" (UID: \"8107394e-a9a2-43a6-bf43-3c785c0455c6\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.192756 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95khn\" (UniqueName: \"kubernetes.io/projected/a1c0aebf-0263-4947-88ee-70a119ef3453-kube-api-access-95khn\") pod \"a1c0aebf-0263-4947-88ee-70a119ef3453\" (UID: \"a1c0aebf-0263-4947-88ee-70a119ef3453\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.192767 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/018d4544-4c86-45b0-a320-77d857f6e49b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "018d4544-4c86-45b0-a320-77d857f6e49b" (UID: "018d4544-4c86-45b0-a320-77d857f6e49b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.194506 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8107394e-a9a2-43a6-bf43-3c785c0455c6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8107394e-a9a2-43a6-bf43-3c785c0455c6" (UID: "8107394e-a9a2-43a6-bf43-3c785c0455c6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.196860 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8107394e-a9a2-43a6-bf43-3c785c0455c6-kube-api-access-z2sk9" (OuterVolumeSpecName: "kube-api-access-z2sk9") pod "8107394e-a9a2-43a6-bf43-3c785c0455c6" (UID: "8107394e-a9a2-43a6-bf43-3c785c0455c6"). InnerVolumeSpecName "kube-api-access-z2sk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.199669 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/018d4544-4c86-45b0-a320-77d857f6e49b-kube-api-access-xhlg7" (OuterVolumeSpecName: "kube-api-access-xhlg7") pod "018d4544-4c86-45b0-a320-77d857f6e49b" (UID: "018d4544-4c86-45b0-a320-77d857f6e49b"). InnerVolumeSpecName "kube-api-access-xhlg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.199725 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85b66a20-d9b6-45ca-889b-87121ebac3f0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.199756 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8107394e-a9a2-43a6-bf43-3c785c0455c6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.199807 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/018d4544-4c86-45b0-a320-77d857f6e49b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.199822 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1c0aebf-0263-4947-88ee-70a119ef3453-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.199834 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6d47cc8-65ed-4b57-abb2-2216c9ed451a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.199850 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvt4t\" (UniqueName: \"kubernetes.io/projected/a6d47cc8-65ed-4b57-abb2-2216c9ed451a-kube-api-access-mvt4t\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.199863 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2sk9\" (UniqueName: \"kubernetes.io/projected/8107394e-a9a2-43a6-bf43-3c785c0455c6-kube-api-access-z2sk9\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.201792 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85b66a20-d9b6-45ca-889b-87121ebac3f0-kube-api-access-t92h4" (OuterVolumeSpecName: "kube-api-access-t92h4") pod "85b66a20-d9b6-45ca-889b-87121ebac3f0" (UID: "85b66a20-d9b6-45ca-889b-87121ebac3f0"). InnerVolumeSpecName "kube-api-access-t92h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.203552 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c0aebf-0263-4947-88ee-70a119ef3453-kube-api-access-95khn" (OuterVolumeSpecName: "kube-api-access-95khn") pod "a1c0aebf-0263-4947-88ee-70a119ef3453" (UID: "a1c0aebf-0263-4947-88ee-70a119ef3453"). InnerVolumeSpecName "kube-api-access-95khn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.300788 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwh26\" (UniqueName: \"kubernetes.io/projected/31e91520-3698-4c0a-82f7-9792c769f9f3-kube-api-access-lwh26\") pod \"31e91520-3698-4c0a-82f7-9792c769f9f3\" (UID: \"31e91520-3698-4c0a-82f7-9792c769f9f3\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.300874 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf8st\" (UniqueName: \"kubernetes.io/projected/87ab7c0d-04a8-4527-966d-d48abfc3c645-kube-api-access-vf8st\") pod \"87ab7c0d-04a8-4527-966d-d48abfc3c645\" (UID: \"87ab7c0d-04a8-4527-966d-d48abfc3c645\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.300981 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87ab7c0d-04a8-4527-966d-d48abfc3c645-operator-scripts\") pod \"87ab7c0d-04a8-4527-966d-d48abfc3c645\" (UID: \"87ab7c0d-04a8-4527-966d-d48abfc3c645\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.301026 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqdld\" (UniqueName: \"kubernetes.io/projected/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec-kube-api-access-dqdld\") pod \"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec\" (UID: \"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.301075 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31e91520-3698-4c0a-82f7-9792c769f9f3-operator-scripts\") pod \"31e91520-3698-4c0a-82f7-9792c769f9f3\" (UID: \"31e91520-3698-4c0a-82f7-9792c769f9f3\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.301153 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec-operator-scripts\") pod \"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec\" (UID: \"5c8567c2-527e-4cba-9d7b-e3da31d2e0ec\") " Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.301775 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87ab7c0d-04a8-4527-966d-d48abfc3c645-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "87ab7c0d-04a8-4527-966d-d48abfc3c645" (UID: "87ab7c0d-04a8-4527-966d-d48abfc3c645"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.301843 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c8567c2-527e-4cba-9d7b-e3da31d2e0ec" (UID: "5c8567c2-527e-4cba-9d7b-e3da31d2e0ec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.302131 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31e91520-3698-4c0a-82f7-9792c769f9f3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "31e91520-3698-4c0a-82f7-9792c769f9f3" (UID: "31e91520-3698-4c0a-82f7-9792c769f9f3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.302205 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87ab7c0d-04a8-4527-966d-d48abfc3c645-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.302223 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95khn\" (UniqueName: \"kubernetes.io/projected/a1c0aebf-0263-4947-88ee-70a119ef3453-kube-api-access-95khn\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.302236 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t92h4\" (UniqueName: \"kubernetes.io/projected/85b66a20-d9b6-45ca-889b-87121ebac3f0-kube-api-access-t92h4\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.302244 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhlg7\" (UniqueName: \"kubernetes.io/projected/018d4544-4c86-45b0-a320-77d857f6e49b-kube-api-access-xhlg7\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.302282 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.307320 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31e91520-3698-4c0a-82f7-9792c769f9f3-kube-api-access-lwh26" (OuterVolumeSpecName: "kube-api-access-lwh26") pod "31e91520-3698-4c0a-82f7-9792c769f9f3" (UID: "31e91520-3698-4c0a-82f7-9792c769f9f3"). InnerVolumeSpecName "kube-api-access-lwh26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.307744 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec-kube-api-access-dqdld" (OuterVolumeSpecName: "kube-api-access-dqdld") pod "5c8567c2-527e-4cba-9d7b-e3da31d2e0ec" (UID: "5c8567c2-527e-4cba-9d7b-e3da31d2e0ec"). InnerVolumeSpecName "kube-api-access-dqdld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.308290 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87ab7c0d-04a8-4527-966d-d48abfc3c645-kube-api-access-vf8st" (OuterVolumeSpecName: "kube-api-access-vf8st") pod "87ab7c0d-04a8-4527-966d-d48abfc3c645" (UID: "87ab7c0d-04a8-4527-966d-d48abfc3c645"). InnerVolumeSpecName "kube-api-access-vf8st". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.404454 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwh26\" (UniqueName: \"kubernetes.io/projected/31e91520-3698-4c0a-82f7-9792c769f9f3-kube-api-access-lwh26\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.404509 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf8st\" (UniqueName: \"kubernetes.io/projected/87ab7c0d-04a8-4527-966d-d48abfc3c645-kube-api-access-vf8st\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.404530 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqdld\" (UniqueName: \"kubernetes.io/projected/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec-kube-api-access-dqdld\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:22 crc kubenswrapper[4857]: I1201 21:54:22.404548 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31e91520-3698-4c0a-82f7-9792c769f9f3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:22.999450 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9gzqb" event={"ID":"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab","Type":"ContainerStarted","Data":"7450f895776482d0d31ffe959649c74c1d508b75293a44a04f13a71ef0ca3ed5"} Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:23.002644 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a859-account-create-update-z4slz" Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:23.002693 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-27ab-account-create-update-4mr2x" Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:23.003168 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"a448d85e73aeb8509929fb3ad12073492702f2602ec66a62d70a874fb9edb1df"} Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:23.003237 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"3dd00d6671d5d95e078f47016df83452fe4cf3f2640119c6e92203e7d38cce55"} Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:23.003259 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"22ad495bb334e8f08c3bbe1b428b734e2ab7eb827b235d2aadc97b2d1f6352e1"} Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:23.003460 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fac7-account-create-update-h6hzj" Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:23.003529 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-qm6hg" Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:23.004314 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v2h7r" Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:23.003468 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zdjss" Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:23.004353 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2fb1-account-create-update-66k42" Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:23.024108 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9k4pr" Dec 01 21:54:23 crc kubenswrapper[4857]: I1201 21:54:23.041932 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-9gzqb" podStartSLOduration=9.367580043 podStartE2EDuration="14.041900081s" podCreationTimestamp="2025-12-01 21:54:09 +0000 UTC" firstStartedPulling="2025-12-01 21:54:17.128555069 +0000 UTC m=+1175.618617386" lastFinishedPulling="2025-12-01 21:54:21.802875107 +0000 UTC m=+1180.292937424" observedRunningTime="2025-12-01 21:54:23.022240425 +0000 UTC m=+1181.512302742" watchObservedRunningTime="2025-12-01 21:54:23.041900081 +0000 UTC m=+1181.531962428" Dec 01 21:54:25 crc kubenswrapper[4857]: I1201 21:54:25.022124 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"3b284b23a52d2110daae91b709ffbaefaf93e0bf98afa9516175d07faaafb715"} Dec 01 21:54:25 crc kubenswrapper[4857]: I1201 21:54:25.023759 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"fb1d8262be143592d4c770a330efa2ccbbf0f66bf9c8395e31e56a44cf202bcd"} Dec 01 21:54:25 crc kubenswrapper[4857]: I1201 21:54:25.023802 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"a2a20f30dc6c479bfc6e809507d725ccd5642626272ae2b4e79bb4633125a8fc"} Dec 01 21:54:25 crc kubenswrapper[4857]: I1201 21:54:25.023815 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"3bf42fa3842b2baaa96bbd5842c9d4b6195beb9889d7cbe4ccdb9506c0d0e33a"} Dec 01 21:54:27 crc kubenswrapper[4857]: I1201 21:54:27.044696 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"2dc48cd4645c1358fa10306fdbd1cee6d36590982c24d5efe77cbb82fb128d7b"} Dec 01 21:54:27 crc kubenswrapper[4857]: I1201 21:54:27.046501 4857 generic.go:334] "Generic (PLEG): container finished" podID="dab7e48a-a3a1-4e9a-bfee-0e0fa712adab" containerID="7450f895776482d0d31ffe959649c74c1d508b75293a44a04f13a71ef0ca3ed5" exitCode=0 Dec 01 21:54:27 crc kubenswrapper[4857]: I1201 21:54:27.046533 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9gzqb" event={"ID":"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab","Type":"ContainerDied","Data":"7450f895776482d0d31ffe959649c74c1d508b75293a44a04f13a71ef0ca3ed5"} Dec 01 21:54:28 crc kubenswrapper[4857]: I1201 21:54:28.065910 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"f0cc97aa047f3486c6ab6afb390ac76132a076bad9958a8018092e40ae987a9c"} Dec 01 21:54:28 crc kubenswrapper[4857]: I1201 21:54:28.066326 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"2b52786a4cbf894ab8e96eb29128a78ea8f69448acfa8d136e0ed631757fa08b"} Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.066740 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.079612 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"b75dbc23d64e9b8a90ada6b4a7c938de1aa2cfdecd948ab2e2f357071eaa0d40"} Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.079663 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"82ef545428fe1aa724482f7fab57d962fb475d990c2c3ab792548e9d85a563ec"} Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.082078 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9f69d" event={"ID":"f985748e-24cd-4d1b-abe8-7a579aee87af","Type":"ContainerStarted","Data":"3ade7b8b1e44a17f41d437ae925738bf97369c91a5e3f1c839100e210b78bbf7"} Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.092823 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9gzqb" event={"ID":"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab","Type":"ContainerDied","Data":"cb9ac852a35faf061273d1f26f8c345b3288bad92344bfd76c134d27ef6e22ab"} Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.092871 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb9ac852a35faf061273d1f26f8c345b3288bad92344bfd76c134d27ef6e22ab" Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.092937 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9gzqb" Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.110104 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-9f69d" podStartSLOduration=2.439835237 podStartE2EDuration="32.110084812s" podCreationTimestamp="2025-12-01 21:53:57 +0000 UTC" firstStartedPulling="2025-12-01 21:53:58.033025701 +0000 UTC m=+1156.523088018" lastFinishedPulling="2025-12-01 21:54:27.703275236 +0000 UTC m=+1186.193337593" observedRunningTime="2025-12-01 21:54:29.105261935 +0000 UTC m=+1187.595324252" watchObservedRunningTime="2025-12-01 21:54:29.110084812 +0000 UTC m=+1187.600147129" Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.184063 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-combined-ca-bundle\") pod \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\" (UID: \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\") " Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.184110 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkmqk\" (UniqueName: \"kubernetes.io/projected/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-kube-api-access-bkmqk\") pod \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\" (UID: \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\") " Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.184157 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-config-data\") pod \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\" (UID: \"dab7e48a-a3a1-4e9a-bfee-0e0fa712adab\") " Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.195819 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-kube-api-access-bkmqk" (OuterVolumeSpecName: "kube-api-access-bkmqk") pod "dab7e48a-a3a1-4e9a-bfee-0e0fa712adab" (UID: "dab7e48a-a3a1-4e9a-bfee-0e0fa712adab"). InnerVolumeSpecName "kube-api-access-bkmqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.221207 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dab7e48a-a3a1-4e9a-bfee-0e0fa712adab" (UID: "dab7e48a-a3a1-4e9a-bfee-0e0fa712adab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.263333 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-config-data" (OuterVolumeSpecName: "config-data") pod "dab7e48a-a3a1-4e9a-bfee-0e0fa712adab" (UID: "dab7e48a-a3a1-4e9a-bfee-0e0fa712adab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.285922 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.285953 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkmqk\" (UniqueName: \"kubernetes.io/projected/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-kube-api-access-bkmqk\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:29 crc kubenswrapper[4857]: I1201 21:54:29.285964 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.116730 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"0b119b33281539ea6c365a11c9803c026ee61a24bd29591944a61a3aa111af97"} Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.117138 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ea80def-31e0-4cc2-9f15-26ee35c62641","Type":"ContainerStarted","Data":"9ec95e980a4758644176f8eca071bb557b49645227c8d25e83ffbd224ed2ca48"} Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.192331 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=46.972433749 podStartE2EDuration="56.192303448s" podCreationTimestamp="2025-12-01 21:53:34 +0000 UTC" firstStartedPulling="2025-12-01 21:54:17.247713225 +0000 UTC m=+1175.737775532" lastFinishedPulling="2025-12-01 21:54:26.467582914 +0000 UTC m=+1184.957645231" observedRunningTime="2025-12-01 21:54:30.185576126 +0000 UTC m=+1188.675638443" watchObservedRunningTime="2025-12-01 21:54:30.192303448 +0000 UTC m=+1188.682365785" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.346725 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-t9vh9"] Dec 01 21:54:30 crc kubenswrapper[4857]: E1201 21:54:30.347097 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab7e48a-a3a1-4e9a-bfee-0e0fa712adab" containerName="keystone-db-sync" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347109 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab7e48a-a3a1-4e9a-bfee-0e0fa712adab" containerName="keystone-db-sync" Dec 01 21:54:30 crc kubenswrapper[4857]: E1201 21:54:30.347131 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31e91520-3698-4c0a-82f7-9792c769f9f3" containerName="mariadb-account-create-update" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347138 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="31e91520-3698-4c0a-82f7-9792c769f9f3" containerName="mariadb-account-create-update" Dec 01 21:54:30 crc kubenswrapper[4857]: E1201 21:54:30.347147 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8107394e-a9a2-43a6-bf43-3c785c0455c6" containerName="mariadb-account-create-update" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347154 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8107394e-a9a2-43a6-bf43-3c785c0455c6" containerName="mariadb-account-create-update" Dec 01 21:54:30 crc kubenswrapper[4857]: E1201 21:54:30.347165 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d47cc8-65ed-4b57-abb2-2216c9ed451a" containerName="mariadb-database-create" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347171 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d47cc8-65ed-4b57-abb2-2216c9ed451a" containerName="mariadb-database-create" Dec 01 21:54:30 crc kubenswrapper[4857]: E1201 21:54:30.347183 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85b66a20-d9b6-45ca-889b-87121ebac3f0" containerName="mariadb-database-create" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347190 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="85b66a20-d9b6-45ca-889b-87121ebac3f0" containerName="mariadb-database-create" Dec 01 21:54:30 crc kubenswrapper[4857]: E1201 21:54:30.347203 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ab7c0d-04a8-4527-966d-d48abfc3c645" containerName="mariadb-account-create-update" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347209 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ab7c0d-04a8-4527-966d-d48abfc3c645" containerName="mariadb-account-create-update" Dec 01 21:54:30 crc kubenswrapper[4857]: E1201 21:54:30.347218 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018d4544-4c86-45b0-a320-77d857f6e49b" containerName="mariadb-database-create" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347224 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="018d4544-4c86-45b0-a320-77d857f6e49b" containerName="mariadb-database-create" Dec 01 21:54:30 crc kubenswrapper[4857]: E1201 21:54:30.347246 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c8567c2-527e-4cba-9d7b-e3da31d2e0ec" containerName="mariadb-database-create" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347252 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c8567c2-527e-4cba-9d7b-e3da31d2e0ec" containerName="mariadb-database-create" Dec 01 21:54:30 crc kubenswrapper[4857]: E1201 21:54:30.347278 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c0aebf-0263-4947-88ee-70a119ef3453" containerName="mariadb-account-create-update" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347286 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c0aebf-0263-4947-88ee-70a119ef3453" containerName="mariadb-account-create-update" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347426 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c8567c2-527e-4cba-9d7b-e3da31d2e0ec" containerName="mariadb-database-create" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347440 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="8107394e-a9a2-43a6-bf43-3c785c0455c6" containerName="mariadb-account-create-update" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347447 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="85b66a20-d9b6-45ca-889b-87121ebac3f0" containerName="mariadb-database-create" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347457 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6d47cc8-65ed-4b57-abb2-2216c9ed451a" containerName="mariadb-database-create" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347478 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="87ab7c0d-04a8-4527-966d-d48abfc3c645" containerName="mariadb-account-create-update" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347488 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c0aebf-0263-4947-88ee-70a119ef3453" containerName="mariadb-account-create-update" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347500 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="31e91520-3698-4c0a-82f7-9792c769f9f3" containerName="mariadb-account-create-update" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347513 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="018d4544-4c86-45b0-a320-77d857f6e49b" containerName="mariadb-database-create" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.347523 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="dab7e48a-a3a1-4e9a-bfee-0e0fa712adab" containerName="keystone-db-sync" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.348064 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.356316 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.356358 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.356453 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.356500 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6ll6v" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.358175 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.363166 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t9vh9"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.385566 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-ft69j"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.393773 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.426845 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-ft69j"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.507955 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-xzd26"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.508976 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xzd26" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.510585 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-config\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.510619 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-config-data\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.510670 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.510690 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.510720 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hjgq\" (UniqueName: \"kubernetes.io/projected/a092473d-f071-4398-be53-a8c7fcd6080d-kube-api-access-2hjgq\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.510735 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-credential-keys\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.510759 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-fernet-keys\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.510776 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59xfn\" (UniqueName: \"kubernetes.io/projected/34afe58f-9fa9-4fc5-86fa-83b086821fac-kube-api-access-59xfn\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.510805 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-scripts\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.510830 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-combined-ca-bundle\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.510850 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-dns-svc\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.514328 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.516140 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-djd7n" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.521695 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-xzd26"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.613276 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-x296d"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.614472 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615047 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgbf7\" (UniqueName: \"kubernetes.io/projected/aa7ace20-c228-447f-98b7-dc09e562d5b4-kube-api-access-dgbf7\") pod \"heat-db-sync-xzd26\" (UID: \"aa7ace20-c228-447f-98b7-dc09e562d5b4\") " pod="openstack/heat-db-sync-xzd26" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615113 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615141 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615175 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hjgq\" (UniqueName: \"kubernetes.io/projected/a092473d-f071-4398-be53-a8c7fcd6080d-kube-api-access-2hjgq\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615191 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-credential-keys\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615208 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7ace20-c228-447f-98b7-dc09e562d5b4-combined-ca-bundle\") pod \"heat-db-sync-xzd26\" (UID: \"aa7ace20-c228-447f-98b7-dc09e562d5b4\") " pod="openstack/heat-db-sync-xzd26" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615233 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-fernet-keys\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615250 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59xfn\" (UniqueName: \"kubernetes.io/projected/34afe58f-9fa9-4fc5-86fa-83b086821fac-kube-api-access-59xfn\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615272 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7ace20-c228-447f-98b7-dc09e562d5b4-config-data\") pod \"heat-db-sync-xzd26\" (UID: \"aa7ace20-c228-447f-98b7-dc09e562d5b4\") " pod="openstack/heat-db-sync-xzd26" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615305 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-scripts\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615332 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-combined-ca-bundle\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615351 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-dns-svc\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615399 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-config\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.615417 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-config-data\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.617187 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.617722 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.624809 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-dns-svc\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.625034 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-config\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.629064 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-pkbkh"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.630206 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.631003 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-d5kl6" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.631142 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.631402 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.632547 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-fernet-keys\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.637433 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-scripts\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.639115 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-combined-ca-bundle\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.643659 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-credential-keys\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.644167 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.644330 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-dd4l5" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.644487 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.644901 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-config-data\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.645089 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-x296d"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.652490 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59xfn\" (UniqueName: \"kubernetes.io/projected/34afe58f-9fa9-4fc5-86fa-83b086821fac-kube-api-access-59xfn\") pod \"keystone-bootstrap-t9vh9\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.665950 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hjgq\" (UniqueName: \"kubernetes.io/projected/a092473d-f071-4398-be53-a8c7fcd6080d-kube-api-access-2hjgq\") pod \"dnsmasq-dns-f877ddd87-ft69j\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.674208 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-ft69j"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.674671 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.674905 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.706523 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-pkbkh"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.717723 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7ace20-c228-447f-98b7-dc09e562d5b4-combined-ca-bundle\") pod \"heat-db-sync-xzd26\" (UID: \"aa7ace20-c228-447f-98b7-dc09e562d5b4\") " pod="openstack/heat-db-sync-xzd26" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.717770 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-scripts\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.717799 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-logs\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.717818 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7ace20-c228-447f-98b7-dc09e562d5b4-config-data\") pod \"heat-db-sync-xzd26\" (UID: \"aa7ace20-c228-447f-98b7-dc09e562d5b4\") " pod="openstack/heat-db-sync-xzd26" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.717841 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dc87da26-cfe1-401c-b115-bab361f3c126-etc-machine-id\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.717872 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-config-data\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.717896 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-scripts\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.717925 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-combined-ca-bundle\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.717957 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w855q\" (UniqueName: \"kubernetes.io/projected/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-kube-api-access-w855q\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.717985 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-combined-ca-bundle\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.718059 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-db-sync-config-data\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.718076 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-config-data\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.718116 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgbf7\" (UniqueName: \"kubernetes.io/projected/aa7ace20-c228-447f-98b7-dc09e562d5b4-kube-api-access-dgbf7\") pod \"heat-db-sync-xzd26\" (UID: \"aa7ace20-c228-447f-98b7-dc09e562d5b4\") " pod="openstack/heat-db-sync-xzd26" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.718147 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6pch\" (UniqueName: \"kubernetes.io/projected/dc87da26-cfe1-401c-b115-bab361f3c126-kube-api-access-r6pch\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.745617 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7ace20-c228-447f-98b7-dc09e562d5b4-combined-ca-bundle\") pod \"heat-db-sync-xzd26\" (UID: \"aa7ace20-c228-447f-98b7-dc09e562d5b4\") " pod="openstack/heat-db-sync-xzd26" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.748979 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7ace20-c228-447f-98b7-dc09e562d5b4-config-data\") pod \"heat-db-sync-xzd26\" (UID: \"aa7ace20-c228-447f-98b7-dc09e562d5b4\") " pod="openstack/heat-db-sync-xzd26" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.752776 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgbf7\" (UniqueName: \"kubernetes.io/projected/aa7ace20-c228-447f-98b7-dc09e562d5b4-kube-api-access-dgbf7\") pod \"heat-db-sync-xzd26\" (UID: \"aa7ace20-c228-447f-98b7-dc09e562d5b4\") " pod="openstack/heat-db-sync-xzd26" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.779089 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-ppkc6"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.780153 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.784485 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.787879 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-894zj" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.841201 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ppkc6"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.841941 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-db-sync-config-data\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.841995 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-config-data\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.842188 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6pch\" (UniqueName: \"kubernetes.io/projected/dc87da26-cfe1-401c-b115-bab361f3c126-kube-api-access-r6pch\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.842875 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-scripts\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.842942 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-logs\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.843134 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dc87da26-cfe1-401c-b115-bab361f3c126-etc-machine-id\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.843183 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-config-data\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.843349 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-scripts\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.843585 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-combined-ca-bundle\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.843705 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w855q\" (UniqueName: \"kubernetes.io/projected/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-kube-api-access-w855q\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.843875 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-combined-ca-bundle\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.849661 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xzd26" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.853698 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dc87da26-cfe1-401c-b115-bab361f3c126-etc-machine-id\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.856325 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-combined-ca-bundle\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.886003 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-db-sync-config-data\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.893945 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-scripts\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.895190 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-config-data\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.897334 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-scripts\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.898297 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-config-data\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.900242 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-logs\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.902898 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-combined-ca-bundle\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.923668 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w855q\" (UniqueName: \"kubernetes.io/projected/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-kube-api-access-w855q\") pod \"placement-db-sync-pkbkh\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.924824 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6pch\" (UniqueName: \"kubernetes.io/projected/dc87da26-cfe1-401c-b115-bab361f3c126-kube-api-access-r6pch\") pod \"cinder-db-sync-x296d\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.932417 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-twrhr"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.949836 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c9aa39-8cad-4807-98c7-ee329d87bb20-combined-ca-bundle\") pod \"barbican-db-sync-ppkc6\" (UID: \"66c9aa39-8cad-4807-98c7-ee329d87bb20\") " pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.950087 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66c9aa39-8cad-4807-98c7-ee329d87bb20-db-sync-config-data\") pod \"barbican-db-sync-ppkc6\" (UID: \"66c9aa39-8cad-4807-98c7-ee329d87bb20\") " pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.950129 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjtrt\" (UniqueName: \"kubernetes.io/projected/66c9aa39-8cad-4807-98c7-ee329d87bb20-kube-api-access-pjtrt\") pod \"barbican-db-sync-ppkc6\" (UID: \"66c9aa39-8cad-4807-98c7-ee329d87bb20\") " pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.976061 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pkbkh" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.996705 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-twrhr"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.996738 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-twrhr"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.996748 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-s58fk"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.997117 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-twrhr" Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.997552 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-s58fk"] Dec 01 21:54:30 crc kubenswrapper[4857]: I1201 21:54:30.997621 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-s58fk" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:30.999838 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-6jgxw" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.000436 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.000823 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.020944 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-k46r8"] Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.022236 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.022914 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.024304 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.030196 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-k46r8"] Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.030380 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-twrhr" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.030503 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.030942 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.031361 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.037866 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.060194 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.060435 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8dc6c548-e991-444e-a627-67896bf3e307-config\") pod \"neutron-db-sync-s58fk\" (UID: \"8dc6c548-e991-444e-a627-67896bf3e307\") " pod="openstack/neutron-db-sync-s58fk" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.060459 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn5xp\" (UniqueName: \"kubernetes.io/projected/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-kube-api-access-sn5xp\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.060504 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ht2s\" (UniqueName: \"kubernetes.io/projected/8dc6c548-e991-444e-a627-67896bf3e307-kube-api-access-9ht2s\") pod \"neutron-db-sync-s58fk\" (UID: \"8dc6c548-e991-444e-a627-67896bf3e307\") " pod="openstack/neutron-db-sync-s58fk" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.060523 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.060551 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.060584 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-config\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.060608 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66c9aa39-8cad-4807-98c7-ee329d87bb20-db-sync-config-data\") pod \"barbican-db-sync-ppkc6\" (UID: \"66c9aa39-8cad-4807-98c7-ee329d87bb20\") " pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.060628 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.060659 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjtrt\" (UniqueName: \"kubernetes.io/projected/66c9aa39-8cad-4807-98c7-ee329d87bb20-kube-api-access-pjtrt\") pod \"barbican-db-sync-ppkc6\" (UID: \"66c9aa39-8cad-4807-98c7-ee329d87bb20\") " pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.060683 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dc6c548-e991-444e-a627-67896bf3e307-combined-ca-bundle\") pod \"neutron-db-sync-s58fk\" (UID: \"8dc6c548-e991-444e-a627-67896bf3e307\") " pod="openstack/neutron-db-sync-s58fk" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.060704 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c9aa39-8cad-4807-98c7-ee329d87bb20-combined-ca-bundle\") pod \"barbican-db-sync-ppkc6\" (UID: \"66c9aa39-8cad-4807-98c7-ee329d87bb20\") " pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.067496 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c9aa39-8cad-4807-98c7-ee329d87bb20-combined-ca-bundle\") pod \"barbican-db-sync-ppkc6\" (UID: \"66c9aa39-8cad-4807-98c7-ee329d87bb20\") " pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.082691 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66c9aa39-8cad-4807-98c7-ee329d87bb20-db-sync-config-data\") pod \"barbican-db-sync-ppkc6\" (UID: \"66c9aa39-8cad-4807-98c7-ee329d87bb20\") " pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.084874 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjtrt\" (UniqueName: \"kubernetes.io/projected/66c9aa39-8cad-4807-98c7-ee329d87bb20-kube-api-access-pjtrt\") pod \"barbican-db-sync-ppkc6\" (UID: \"66c9aa39-8cad-4807-98c7-ee329d87bb20\") " pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.127832 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-x296d" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.127872 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-twrhr" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.161762 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.161823 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.161850 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8dc6c548-e991-444e-a627-67896bf3e307-config\") pod \"neutron-db-sync-s58fk\" (UID: \"8dc6c548-e991-444e-a627-67896bf3e307\") " pod="openstack/neutron-db-sync-s58fk" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.161953 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn5xp\" (UniqueName: \"kubernetes.io/projected/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-kube-api-access-sn5xp\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.162074 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ht2s\" (UniqueName: \"kubernetes.io/projected/8dc6c548-e991-444e-a627-67896bf3e307-kube-api-access-9ht2s\") pod \"neutron-db-sync-s58fk\" (UID: \"8dc6c548-e991-444e-a627-67896bf3e307\") " pod="openstack/neutron-db-sync-s58fk" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.162113 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.162135 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.162239 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f166659f-d3ef-4f73-8258-46f19eacd6d2-run-httpd\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.162259 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-config\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.162285 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-scripts\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.162340 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.162363 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2pkf\" (UniqueName: \"kubernetes.io/projected/f166659f-d3ef-4f73-8258-46f19eacd6d2-kube-api-access-b2pkf\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.162384 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f166659f-d3ef-4f73-8258-46f19eacd6d2-log-httpd\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.162413 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.162454 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dc6c548-e991-444e-a627-67896bf3e307-combined-ca-bundle\") pod \"neutron-db-sync-s58fk\" (UID: \"8dc6c548-e991-444e-a627-67896bf3e307\") " pod="openstack/neutron-db-sync-s58fk" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.162486 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-config-data\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.163972 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.164765 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.166692 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.168543 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-config\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.168718 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.185494 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8dc6c548-e991-444e-a627-67896bf3e307-config\") pod \"neutron-db-sync-s58fk\" (UID: \"8dc6c548-e991-444e-a627-67896bf3e307\") " pod="openstack/neutron-db-sync-s58fk" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.187290 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dc6c548-e991-444e-a627-67896bf3e307-combined-ca-bundle\") pod \"neutron-db-sync-s58fk\" (UID: \"8dc6c548-e991-444e-a627-67896bf3e307\") " pod="openstack/neutron-db-sync-s58fk" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.192387 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn5xp\" (UniqueName: \"kubernetes.io/projected/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-kube-api-access-sn5xp\") pod \"dnsmasq-dns-58dd9ff6bc-k46r8\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.197204 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ht2s\" (UniqueName: \"kubernetes.io/projected/8dc6c548-e991-444e-a627-67896bf3e307-kube-api-access-9ht2s\") pod \"neutron-db-sync-s58fk\" (UID: \"8dc6c548-e991-444e-a627-67896bf3e307\") " pod="openstack/neutron-db-sync-s58fk" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.257016 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-twrhr"] Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.264149 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2pkf\" (UniqueName: \"kubernetes.io/projected/f166659f-d3ef-4f73-8258-46f19eacd6d2-kube-api-access-b2pkf\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.264204 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f166659f-d3ef-4f73-8258-46f19eacd6d2-log-httpd\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.264230 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.264261 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-config-data\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.264321 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.264417 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f166659f-d3ef-4f73-8258-46f19eacd6d2-run-httpd\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.264442 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-scripts\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.264911 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f166659f-d3ef-4f73-8258-46f19eacd6d2-log-httpd\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.265231 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f166659f-d3ef-4f73-8258-46f19eacd6d2-run-httpd\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.267150 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-twrhr"] Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.291967 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.293360 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.295811 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2pkf\" (UniqueName: \"kubernetes.io/projected/f166659f-d3ef-4f73-8258-46f19eacd6d2-kube-api-access-b2pkf\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.297275 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-config-data\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.304898 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-scripts\") pod \"ceilometer-0\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.310289 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.377498 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-s58fk" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.407114 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.509438 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.586101 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-ft69j"] Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.793117 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t9vh9"] Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.906281 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-xzd26"] Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.939832 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-pkbkh"] Dec 01 21:54:31 crc kubenswrapper[4857]: I1201 21:54:31.981439 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-x296d"] Dec 01 21:54:32 crc kubenswrapper[4857]: I1201 21:54:32.163300 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-ft69j" event={"ID":"a092473d-f071-4398-be53-a8c7fcd6080d","Type":"ContainerStarted","Data":"3c05e3ed9b8d83c7b9d7736349ae999faa73237b7b504f0ed1705ae7f8960a29"} Dec 01 21:54:32 crc kubenswrapper[4857]: I1201 21:54:32.163364 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-ft69j" event={"ID":"a092473d-f071-4398-be53-a8c7fcd6080d","Type":"ContainerStarted","Data":"34c45c07879ea1633c2a30395569dea64fd6a7d75b71acf92b0f569ecd05297e"} Dec 01 21:54:32 crc kubenswrapper[4857]: I1201 21:54:32.163550 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f877ddd87-ft69j" podUID="a092473d-f071-4398-be53-a8c7fcd6080d" containerName="init" containerID="cri-o://3c05e3ed9b8d83c7b9d7736349ae999faa73237b7b504f0ed1705ae7f8960a29" gracePeriod=10 Dec 01 21:54:32 crc kubenswrapper[4857]: I1201 21:54:32.177328 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xzd26" event={"ID":"aa7ace20-c228-447f-98b7-dc09e562d5b4","Type":"ContainerStarted","Data":"b7095a0b24874e965f6c5c7d07f7cca9c9e975f6f81cef565d89ef17993b5375"} Dec 01 21:54:32 crc kubenswrapper[4857]: I1201 21:54:32.178784 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-x296d" event={"ID":"dc87da26-cfe1-401c-b115-bab361f3c126","Type":"ContainerStarted","Data":"68e1a6ce912af12a209d3643e35a8fe97485121f4f423cf151ac4c9cbc829c05"} Dec 01 21:54:32 crc kubenswrapper[4857]: I1201 21:54:32.179774 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pkbkh" event={"ID":"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b","Type":"ContainerStarted","Data":"5eab6ed74e50984bb40395e798546547197ae3265f3d46d7da50fb8c37f9cf0d"} Dec 01 21:54:32 crc kubenswrapper[4857]: I1201 21:54:32.180652 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t9vh9" event={"ID":"34afe58f-9fa9-4fc5-86fa-83b086821fac","Type":"ContainerStarted","Data":"9f004838d27d8658b520b713c816cff7b0d67fbe6ed433e10580bef3fad2a5d4"} Dec 01 21:54:32 crc kubenswrapper[4857]: I1201 21:54:32.207663 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ppkc6"] Dec 01 21:54:32 crc kubenswrapper[4857]: I1201 21:54:32.282358 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-s58fk"] Dec 01 21:54:32 crc kubenswrapper[4857]: I1201 21:54:32.375401 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-k46r8"] Dec 01 21:54:32 crc kubenswrapper[4857]: I1201 21:54:32.562844 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.241966 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-s58fk" event={"ID":"8dc6c548-e991-444e-a627-67896bf3e307","Type":"ContainerStarted","Data":"abf1b62202614efcf7bfea038b20ef4221705a3afe75a406ca46b3134cca079a"} Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.242398 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-s58fk" event={"ID":"8dc6c548-e991-444e-a627-67896bf3e307","Type":"ContainerStarted","Data":"41fbc7226100123d3f5439293d825e540538c03d4376f293d795de38e0b1d2ac"} Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.257975 4857 generic.go:334] "Generic (PLEG): container finished" podID="a092473d-f071-4398-be53-a8c7fcd6080d" containerID="3c05e3ed9b8d83c7b9d7736349ae999faa73237b7b504f0ed1705ae7f8960a29" exitCode=0 Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.258052 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-ft69j" event={"ID":"a092473d-f071-4398-be53-a8c7fcd6080d","Type":"ContainerDied","Data":"3c05e3ed9b8d83c7b9d7736349ae999faa73237b7b504f0ed1705ae7f8960a29"} Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.261258 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ppkc6" event={"ID":"66c9aa39-8cad-4807-98c7-ee329d87bb20","Type":"ContainerStarted","Data":"1a00bbd0dd2057d6365d7bd4d4fb27cd9a597b96554336bde7748886ef43929e"} Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.279906 4857 generic.go:334] "Generic (PLEG): container finished" podID="6c2f464a-9080-43f7-bf3e-ac92a9aaea77" containerID="444cb8be53d41af1d94de563163e451ef5fe4e936480f3c85a517a5ee37df0c4" exitCode=0 Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.279976 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" event={"ID":"6c2f464a-9080-43f7-bf3e-ac92a9aaea77","Type":"ContainerDied","Data":"444cb8be53d41af1d94de563163e451ef5fe4e936480f3c85a517a5ee37df0c4"} Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.280006 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" event={"ID":"6c2f464a-9080-43f7-bf3e-ac92a9aaea77","Type":"ContainerStarted","Data":"c042212b472a2246dda7177757dab2916e74011fff85fbacbf92da761523f23a"} Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.304057 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-s58fk" podStartSLOduration=3.303970588 podStartE2EDuration="3.303970588s" podCreationTimestamp="2025-12-01 21:54:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:54:33.267648838 +0000 UTC m=+1191.757711155" watchObservedRunningTime="2025-12-01 21:54:33.303970588 +0000 UTC m=+1191.794032905" Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.312433 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t9vh9" event={"ID":"34afe58f-9fa9-4fc5-86fa-83b086821fac","Type":"ContainerStarted","Data":"c15dd48459e3aae6a3a128fbaee6c013ca37c12ae6d68aaa11e2fa2339a4f126"} Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.405686 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f166659f-d3ef-4f73-8258-46f19eacd6d2","Type":"ContainerStarted","Data":"3647ba61d437efb6da3e37f975093fc94522f182b8150ac50e7f2c060fa3c31e"} Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.433492 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-t9vh9" podStartSLOduration=3.4334715239999998 podStartE2EDuration="3.433471524s" podCreationTimestamp="2025-12-01 21:54:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:54:33.406525241 +0000 UTC m=+1191.896587558" watchObservedRunningTime="2025-12-01 21:54:33.433471524 +0000 UTC m=+1191.923533841" Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.443078 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.559417 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.655736 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-ovsdbserver-sb\") pod \"a092473d-f071-4398-be53-a8c7fcd6080d\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.655840 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-config\") pod \"a092473d-f071-4398-be53-a8c7fcd6080d\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.655918 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-dns-svc\") pod \"a092473d-f071-4398-be53-a8c7fcd6080d\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.656030 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hjgq\" (UniqueName: \"kubernetes.io/projected/a092473d-f071-4398-be53-a8c7fcd6080d-kube-api-access-2hjgq\") pod \"a092473d-f071-4398-be53-a8c7fcd6080d\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.656084 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-ovsdbserver-nb\") pod \"a092473d-f071-4398-be53-a8c7fcd6080d\" (UID: \"a092473d-f071-4398-be53-a8c7fcd6080d\") " Dec 01 21:54:33 crc kubenswrapper[4857]: I1201 21:54:33.982218 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a092473d-f071-4398-be53-a8c7fcd6080d-kube-api-access-2hjgq" (OuterVolumeSpecName: "kube-api-access-2hjgq") pod "a092473d-f071-4398-be53-a8c7fcd6080d" (UID: "a092473d-f071-4398-be53-a8c7fcd6080d"). InnerVolumeSpecName "kube-api-access-2hjgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.004631 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a092473d-f071-4398-be53-a8c7fcd6080d" (UID: "a092473d-f071-4398-be53-a8c7fcd6080d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.004680 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a092473d-f071-4398-be53-a8c7fcd6080d" (UID: "a092473d-f071-4398-be53-a8c7fcd6080d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.005482 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a092473d-f071-4398-be53-a8c7fcd6080d" (UID: "a092473d-f071-4398-be53-a8c7fcd6080d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.041875 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-config" (OuterVolumeSpecName: "config") pod "a092473d-f071-4398-be53-a8c7fcd6080d" (UID: "a092473d-f071-4398-be53-a8c7fcd6080d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.082155 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.082205 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.082215 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.082224 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hjgq\" (UniqueName: \"kubernetes.io/projected/a092473d-f071-4398-be53-a8c7fcd6080d-kube-api-access-2hjgq\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.082233 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a092473d-f071-4398-be53-a8c7fcd6080d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.428183 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-ft69j" event={"ID":"a092473d-f071-4398-be53-a8c7fcd6080d","Type":"ContainerDied","Data":"34c45c07879ea1633c2a30395569dea64fd6a7d75b71acf92b0f569ecd05297e"} Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.428237 4857 scope.go:117] "RemoveContainer" containerID="3c05e3ed9b8d83c7b9d7736349ae999faa73237b7b504f0ed1705ae7f8960a29" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.428366 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-ft69j" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.445131 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" event={"ID":"6c2f464a-9080-43f7-bf3e-ac92a9aaea77","Type":"ContainerStarted","Data":"7cb10cb5343476af2d24bdb3e5c30494885537c087187ee67ea2d82bd9a98c52"} Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.445529 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.562378 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-ft69j"] Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.581537 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-ft69j"] Dec 01 21:54:34 crc kubenswrapper[4857]: I1201 21:54:34.582515 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" podStartSLOduration=4.582504558 podStartE2EDuration="4.582504558s" podCreationTimestamp="2025-12-01 21:54:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:54:34.49961064 +0000 UTC m=+1192.989672957" watchObservedRunningTime="2025-12-01 21:54:34.582504558 +0000 UTC m=+1193.072566875" Dec 01 21:54:35 crc kubenswrapper[4857]: I1201 21:54:35.939715 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a092473d-f071-4398-be53-a8c7fcd6080d" path="/var/lib/kubelet/pods/a092473d-f071-4398-be53-a8c7fcd6080d/volumes" Dec 01 21:54:41 crc kubenswrapper[4857]: I1201 21:54:41.409333 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:54:41 crc kubenswrapper[4857]: I1201 21:54:41.483425 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tcxjl"] Dec 01 21:54:41 crc kubenswrapper[4857]: I1201 21:54:41.483975 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-tcxjl" podUID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerName="dnsmasq-dns" containerID="cri-o://243f268abfc250a13a95a74cf5150ef217a473f49764d58f80409ac94ca5c64c" gracePeriod=10 Dec 01 21:54:41 crc kubenswrapper[4857]: E1201 21:54:41.667688 4857 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod912999c7_66ef_46d7_a7ea_a8bf7b0a42a0.slice/crio-243f268abfc250a13a95a74cf5150ef217a473f49764d58f80409ac94ca5c64c.scope\": RecentStats: unable to find data in memory cache]" Dec 01 21:54:42 crc kubenswrapper[4857]: I1201 21:54:42.536660 4857 generic.go:334] "Generic (PLEG): container finished" podID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerID="243f268abfc250a13a95a74cf5150ef217a473f49764d58f80409ac94ca5c64c" exitCode=0 Dec 01 21:54:42 crc kubenswrapper[4857]: I1201 21:54:42.536896 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tcxjl" event={"ID":"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0","Type":"ContainerDied","Data":"243f268abfc250a13a95a74cf5150ef217a473f49764d58f80409ac94ca5c64c"} Dec 01 21:54:44 crc kubenswrapper[4857]: I1201 21:54:44.567171 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-tcxjl" podUID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Dec 01 21:54:49 crc kubenswrapper[4857]: I1201 21:54:49.567569 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-tcxjl" podUID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Dec 01 21:54:52 crc kubenswrapper[4857]: I1201 21:54:52.655518 4857 generic.go:334] "Generic (PLEG): container finished" podID="34afe58f-9fa9-4fc5-86fa-83b086821fac" containerID="c15dd48459e3aae6a3a128fbaee6c013ca37c12ae6d68aaa11e2fa2339a4f126" exitCode=0 Dec 01 21:54:52 crc kubenswrapper[4857]: I1201 21:54:52.656109 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t9vh9" event={"ID":"34afe58f-9fa9-4fc5-86fa-83b086821fac","Type":"ContainerDied","Data":"c15dd48459e3aae6a3a128fbaee6c013ca37c12ae6d68aaa11e2fa2339a4f126"} Dec 01 21:54:57 crc kubenswrapper[4857]: I1201 21:54:57.806342 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:54:57 crc kubenswrapper[4857]: I1201 21:54:57.809390 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:54:59 crc kubenswrapper[4857]: I1201 21:54:59.572555 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-tcxjl" podUID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Dec 01 21:54:59 crc kubenswrapper[4857]: I1201 21:54:59.572946 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:55:04 crc kubenswrapper[4857]: I1201 21:55:04.573635 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-tcxjl" podUID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Dec 01 21:55:07 crc kubenswrapper[4857]: E1201 21:55:07.313671 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 01 21:55:07 crc kubenswrapper[4857]: E1201 21:55:07.314078 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pjtrt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-ppkc6_openstack(66c9aa39-8cad-4807-98c7-ee329d87bb20): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:55:07 crc kubenswrapper[4857]: E1201 21:55:07.315837 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-ppkc6" podUID="66c9aa39-8cad-4807-98c7-ee329d87bb20" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.420820 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.432865 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.516679 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-scripts\") pod \"34afe58f-9fa9-4fc5-86fa-83b086821fac\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.516725 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-config\") pod \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.516750 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-ovsdbserver-nb\") pod \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.516788 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bnh8\" (UniqueName: \"kubernetes.io/projected/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-kube-api-access-6bnh8\") pod \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.516840 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-ovsdbserver-sb\") pod \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.516870 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-config-data\") pod \"34afe58f-9fa9-4fc5-86fa-83b086821fac\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.517606 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-combined-ca-bundle\") pod \"34afe58f-9fa9-4fc5-86fa-83b086821fac\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.517632 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-dns-svc\") pod \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\" (UID: \"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0\") " Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.517693 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-fernet-keys\") pod \"34afe58f-9fa9-4fc5-86fa-83b086821fac\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.518336 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-credential-keys\") pod \"34afe58f-9fa9-4fc5-86fa-83b086821fac\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.518451 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59xfn\" (UniqueName: \"kubernetes.io/projected/34afe58f-9fa9-4fc5-86fa-83b086821fac-kube-api-access-59xfn\") pod \"34afe58f-9fa9-4fc5-86fa-83b086821fac\" (UID: \"34afe58f-9fa9-4fc5-86fa-83b086821fac\") " Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.523806 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34afe58f-9fa9-4fc5-86fa-83b086821fac-kube-api-access-59xfn" (OuterVolumeSpecName: "kube-api-access-59xfn") pod "34afe58f-9fa9-4fc5-86fa-83b086821fac" (UID: "34afe58f-9fa9-4fc5-86fa-83b086821fac"). InnerVolumeSpecName "kube-api-access-59xfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.523863 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-scripts" (OuterVolumeSpecName: "scripts") pod "34afe58f-9fa9-4fc5-86fa-83b086821fac" (UID: "34afe58f-9fa9-4fc5-86fa-83b086821fac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.525625 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-kube-api-access-6bnh8" (OuterVolumeSpecName: "kube-api-access-6bnh8") pod "912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" (UID: "912999c7-66ef-46d7-a7ea-a8bf7b0a42a0"). InnerVolumeSpecName "kube-api-access-6bnh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.538918 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "34afe58f-9fa9-4fc5-86fa-83b086821fac" (UID: "34afe58f-9fa9-4fc5-86fa-83b086821fac"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.546683 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "34afe58f-9fa9-4fc5-86fa-83b086821fac" (UID: "34afe58f-9fa9-4fc5-86fa-83b086821fac"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.553288 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-config-data" (OuterVolumeSpecName: "config-data") pod "34afe58f-9fa9-4fc5-86fa-83b086821fac" (UID: "34afe58f-9fa9-4fc5-86fa-83b086821fac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.556568 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34afe58f-9fa9-4fc5-86fa-83b086821fac" (UID: "34afe58f-9fa9-4fc5-86fa-83b086821fac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.579059 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" (UID: "912999c7-66ef-46d7-a7ea-a8bf7b0a42a0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.579116 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-config" (OuterVolumeSpecName: "config") pod "912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" (UID: "912999c7-66ef-46d7-a7ea-a8bf7b0a42a0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.581961 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" (UID: "912999c7-66ef-46d7-a7ea-a8bf7b0a42a0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.587814 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" (UID: "912999c7-66ef-46d7-a7ea-a8bf7b0a42a0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.621952 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.621997 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.622014 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.622029 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bnh8\" (UniqueName: \"kubernetes.io/projected/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-kube-api-access-6bnh8\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.622056 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.622067 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.622078 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.622092 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.622102 4857 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.622111 4857 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34afe58f-9fa9-4fc5-86fa-83b086821fac-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.622124 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59xfn\" (UniqueName: \"kubernetes.io/projected/34afe58f-9fa9-4fc5-86fa-83b086821fac-kube-api-access-59xfn\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:07 crc kubenswrapper[4857]: E1201 21:55:07.797569 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 01 21:55:07 crc kubenswrapper[4857]: E1201 21:55:07.797761 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n56h5b9h5f7h97h4h87h5b8h54bh57fhffh96h59ch67h5d6hdbh564h59fh96h88h8fhf6h5b6h5f9h6h5d9h58h98h65fhdhf4h659h694q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b2pkf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(f166659f-d3ef-4f73-8258-46f19eacd6d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.810598 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t9vh9" event={"ID":"34afe58f-9fa9-4fc5-86fa-83b086821fac","Type":"ContainerDied","Data":"9f004838d27d8658b520b713c816cff7b0d67fbe6ed433e10580bef3fad2a5d4"} Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.810640 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t9vh9" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.810668 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f004838d27d8658b520b713c816cff7b0d67fbe6ed433e10580bef3fad2a5d4" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.815116 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tcxjl" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.815167 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tcxjl" event={"ID":"912999c7-66ef-46d7-a7ea-a8bf7b0a42a0","Type":"ContainerDied","Data":"647237fa82377c40370d08596511110ba06e34acdca35bf7ed18da72773b2cc5"} Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.815266 4857 scope.go:117] "RemoveContainer" containerID="243f268abfc250a13a95a74cf5150ef217a473f49764d58f80409ac94ca5c64c" Dec 01 21:55:07 crc kubenswrapper[4857]: E1201 21:55:07.820347 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-ppkc6" podUID="66c9aa39-8cad-4807-98c7-ee329d87bb20" Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.869215 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tcxjl"] Dec 01 21:55:07 crc kubenswrapper[4857]: I1201 21:55:07.878437 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tcxjl"] Dec 01 21:55:08 crc kubenswrapper[4857]: E1201 21:55:08.080258 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Dec 01 21:55:08 crc kubenswrapper[4857]: E1201 21:55:08.080438 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dgbf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-xzd26_openstack(aa7ace20-c228-447f-98b7-dc09e562d5b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:55:08 crc kubenswrapper[4857]: E1201 21:55:08.081641 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-xzd26" podUID="aa7ace20-c228-447f-98b7-dc09e562d5b4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.644776 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-t9vh9"] Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.653273 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-t9vh9"] Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.754867 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nq6w4"] Dec 01 21:55:08 crc kubenswrapper[4857]: E1201 21:55:08.755549 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34afe58f-9fa9-4fc5-86fa-83b086821fac" containerName="keystone-bootstrap" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.755598 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="34afe58f-9fa9-4fc5-86fa-83b086821fac" containerName="keystone-bootstrap" Dec 01 21:55:08 crc kubenswrapper[4857]: E1201 21:55:08.755620 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerName="init" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.755630 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerName="init" Dec 01 21:55:08 crc kubenswrapper[4857]: E1201 21:55:08.755655 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerName="dnsmasq-dns" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.755692 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerName="dnsmasq-dns" Dec 01 21:55:08 crc kubenswrapper[4857]: E1201 21:55:08.755706 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a092473d-f071-4398-be53-a8c7fcd6080d" containerName="init" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.755713 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a092473d-f071-4398-be53-a8c7fcd6080d" containerName="init" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.756103 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="34afe58f-9fa9-4fc5-86fa-83b086821fac" containerName="keystone-bootstrap" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.756153 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerName="dnsmasq-dns" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.756168 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="a092473d-f071-4398-be53-a8c7fcd6080d" containerName="init" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.757196 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.759461 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.759670 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.760101 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.760193 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6ll6v" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.761353 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.769008 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nq6w4"] Dec 01 21:55:08 crc kubenswrapper[4857]: E1201 21:55:08.825624 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-xzd26" podUID="aa7ace20-c228-447f-98b7-dc09e562d5b4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.847269 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-config-data\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.847328 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-combined-ca-bundle\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.847404 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-credential-keys\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.847437 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-scripts\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.847561 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg7z6\" (UniqueName: \"kubernetes.io/projected/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-kube-api-access-dg7z6\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.847612 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-fernet-keys\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.950077 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-credential-keys\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.950146 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-scripts\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.950253 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg7z6\" (UniqueName: \"kubernetes.io/projected/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-kube-api-access-dg7z6\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.950307 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-fernet-keys\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.950394 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-config-data\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.950438 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-combined-ca-bundle\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.955697 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-config-data\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.958335 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-fernet-keys\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.958780 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-credential-keys\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.960821 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-combined-ca-bundle\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.972116 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-scripts\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:08 crc kubenswrapper[4857]: I1201 21:55:08.974009 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg7z6\" (UniqueName: \"kubernetes.io/projected/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-kube-api-access-dg7z6\") pod \"keystone-bootstrap-nq6w4\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:09 crc kubenswrapper[4857]: I1201 21:55:09.091257 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:09 crc kubenswrapper[4857]: I1201 21:55:09.262133 4857 scope.go:117] "RemoveContainer" containerID="eadafa0db16cde34081bb8dd379cae4f4e309967f44216a5d4139184f11a3042" Dec 01 21:55:09 crc kubenswrapper[4857]: E1201 21:55:09.268860 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 01 21:55:09 crc kubenswrapper[4857]: E1201 21:55:09.269120 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r6pch,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-x296d_openstack(dc87da26-cfe1-401c-b115-bab361f3c126): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 21:55:09 crc kubenswrapper[4857]: E1201 21:55:09.270513 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-x296d" podUID="dc87da26-cfe1-401c-b115-bab361f3c126" Dec 01 21:55:09 crc kubenswrapper[4857]: I1201 21:55:09.575583 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-tcxjl" podUID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Dec 01 21:55:09 crc kubenswrapper[4857]: I1201 21:55:09.615062 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nq6w4"] Dec 01 21:55:09 crc kubenswrapper[4857]: I1201 21:55:09.844500 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34afe58f-9fa9-4fc5-86fa-83b086821fac" path="/var/lib/kubelet/pods/34afe58f-9fa9-4fc5-86fa-83b086821fac/volumes" Dec 01 21:55:09 crc kubenswrapper[4857]: I1201 21:55:09.845323 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="912999c7-66ef-46d7-a7ea-a8bf7b0a42a0" path="/var/lib/kubelet/pods/912999c7-66ef-46d7-a7ea-a8bf7b0a42a0/volumes" Dec 01 21:55:09 crc kubenswrapper[4857]: I1201 21:55:09.846137 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nq6w4" event={"ID":"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e","Type":"ContainerStarted","Data":"3130f0155ff3566e47ed83d1be6720ca8282f869599701034d163346b01391fd"} Dec 01 21:55:09 crc kubenswrapper[4857]: I1201 21:55:09.846166 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pkbkh" event={"ID":"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b","Type":"ContainerStarted","Data":"dd0faf071d58c6a335f99b92b025df88c87a5dba8a2a7200805f355393b1b86f"} Dec 01 21:55:10 crc kubenswrapper[4857]: E1201 21:55:09.847300 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-x296d" podUID="dc87da26-cfe1-401c-b115-bab361f3c126" Dec 01 21:55:10 crc kubenswrapper[4857]: I1201 21:55:09.905597 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-pkbkh" podStartSLOduration=3.821175892 podStartE2EDuration="39.905573196s" podCreationTimestamp="2025-12-01 21:54:30 +0000 UTC" firstStartedPulling="2025-12-01 21:54:31.956312714 +0000 UTC m=+1190.446375031" lastFinishedPulling="2025-12-01 21:55:08.040709998 +0000 UTC m=+1226.530772335" observedRunningTime="2025-12-01 21:55:09.899315384 +0000 UTC m=+1228.389377721" watchObservedRunningTime="2025-12-01 21:55:09.905573196 +0000 UTC m=+1228.395635513" Dec 01 21:55:10 crc kubenswrapper[4857]: I1201 21:55:10.851562 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nq6w4" event={"ID":"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e","Type":"ContainerStarted","Data":"163bcfe0cfa5f6ce6ad5cec5512dc20820896e37abca9bba3bb2273e44a3730f"} Dec 01 21:55:10 crc kubenswrapper[4857]: I1201 21:55:10.853566 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f166659f-d3ef-4f73-8258-46f19eacd6d2","Type":"ContainerStarted","Data":"d41567b7fae5b528ba83d21d75265a44c2e44823455abdf64fc298e6d51f7a33"} Dec 01 21:55:10 crc kubenswrapper[4857]: I1201 21:55:10.858151 4857 generic.go:334] "Generic (PLEG): container finished" podID="f985748e-24cd-4d1b-abe8-7a579aee87af" containerID="3ade7b8b1e44a17f41d437ae925738bf97369c91a5e3f1c839100e210b78bbf7" exitCode=0 Dec 01 21:55:10 crc kubenswrapper[4857]: I1201 21:55:10.859120 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9f69d" event={"ID":"f985748e-24cd-4d1b-abe8-7a579aee87af","Type":"ContainerDied","Data":"3ade7b8b1e44a17f41d437ae925738bf97369c91a5e3f1c839100e210b78bbf7"} Dec 01 21:55:10 crc kubenswrapper[4857]: I1201 21:55:10.881190 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nq6w4" podStartSLOduration=2.88116608 podStartE2EDuration="2.88116608s" podCreationTimestamp="2025-12-01 21:55:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:10.8708382 +0000 UTC m=+1229.360900537" watchObservedRunningTime="2025-12-01 21:55:10.88116608 +0000 UTC m=+1229.371228417" Dec 01 21:55:11 crc kubenswrapper[4857]: I1201 21:55:11.872147 4857 generic.go:334] "Generic (PLEG): container finished" podID="30bec9e1-3c58-400b-86fe-0ddcdde2bf8b" containerID="dd0faf071d58c6a335f99b92b025df88c87a5dba8a2a7200805f355393b1b86f" exitCode=0 Dec 01 21:55:11 crc kubenswrapper[4857]: I1201 21:55:11.872267 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pkbkh" event={"ID":"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b","Type":"ContainerDied","Data":"dd0faf071d58c6a335f99b92b025df88c87a5dba8a2a7200805f355393b1b86f"} Dec 01 21:55:14 crc kubenswrapper[4857]: I1201 21:55:14.905421 4857 generic.go:334] "Generic (PLEG): container finished" podID="54a6f0ca-e3e9-47df-bca6-3e4287d5c32e" containerID="163bcfe0cfa5f6ce6ad5cec5512dc20820896e37abca9bba3bb2273e44a3730f" exitCode=0 Dec 01 21:55:14 crc kubenswrapper[4857]: I1201 21:55:14.905517 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nq6w4" event={"ID":"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e","Type":"ContainerDied","Data":"163bcfe0cfa5f6ce6ad5cec5512dc20820896e37abca9bba3bb2273e44a3730f"} Dec 01 21:55:15 crc kubenswrapper[4857]: I1201 21:55:15.917908 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9f69d" event={"ID":"f985748e-24cd-4d1b-abe8-7a579aee87af","Type":"ContainerDied","Data":"9b28193c25aa92a56c9a52ba9527e67c2240be38285a977841a136cf67843917"} Dec 01 21:55:15 crc kubenswrapper[4857]: I1201 21:55:15.917961 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b28193c25aa92a56c9a52ba9527e67c2240be38285a977841a136cf67843917" Dec 01 21:55:15 crc kubenswrapper[4857]: I1201 21:55:15.920730 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pkbkh" event={"ID":"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b","Type":"ContainerDied","Data":"5eab6ed74e50984bb40395e798546547197ae3265f3d46d7da50fb8c37f9cf0d"} Dec 01 21:55:15 crc kubenswrapper[4857]: I1201 21:55:15.920949 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eab6ed74e50984bb40395e798546547197ae3265f3d46d7da50fb8c37f9cf0d" Dec 01 21:55:15 crc kubenswrapper[4857]: I1201 21:55:15.944874 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9f69d" Dec 01 21:55:15 crc kubenswrapper[4857]: I1201 21:55:15.957800 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pkbkh" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.094423 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-config-data\") pod \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.095049 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-scripts\") pod \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.095143 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w855q\" (UniqueName: \"kubernetes.io/projected/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-kube-api-access-w855q\") pod \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.095221 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-combined-ca-bundle\") pod \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.095255 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-logs\") pod \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\" (UID: \"30bec9e1-3c58-400b-86fe-0ddcdde2bf8b\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.095367 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-db-sync-config-data\") pod \"f985748e-24cd-4d1b-abe8-7a579aee87af\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.095597 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6tbd\" (UniqueName: \"kubernetes.io/projected/f985748e-24cd-4d1b-abe8-7a579aee87af-kube-api-access-d6tbd\") pod \"f985748e-24cd-4d1b-abe8-7a579aee87af\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.095625 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-config-data\") pod \"f985748e-24cd-4d1b-abe8-7a579aee87af\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.095684 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-combined-ca-bundle\") pod \"f985748e-24cd-4d1b-abe8-7a579aee87af\" (UID: \"f985748e-24cd-4d1b-abe8-7a579aee87af\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.099291 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-logs" (OuterVolumeSpecName: "logs") pod "30bec9e1-3c58-400b-86fe-0ddcdde2bf8b" (UID: "30bec9e1-3c58-400b-86fe-0ddcdde2bf8b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.103036 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-scripts" (OuterVolumeSpecName: "scripts") pod "30bec9e1-3c58-400b-86fe-0ddcdde2bf8b" (UID: "30bec9e1-3c58-400b-86fe-0ddcdde2bf8b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.105234 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f985748e-24cd-4d1b-abe8-7a579aee87af-kube-api-access-d6tbd" (OuterVolumeSpecName: "kube-api-access-d6tbd") pod "f985748e-24cd-4d1b-abe8-7a579aee87af" (UID: "f985748e-24cd-4d1b-abe8-7a579aee87af"). InnerVolumeSpecName "kube-api-access-d6tbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.109560 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f985748e-24cd-4d1b-abe8-7a579aee87af" (UID: "f985748e-24cd-4d1b-abe8-7a579aee87af"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.129207 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30bec9e1-3c58-400b-86fe-0ddcdde2bf8b" (UID: "30bec9e1-3c58-400b-86fe-0ddcdde2bf8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.155236 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-kube-api-access-w855q" (OuterVolumeSpecName: "kube-api-access-w855q") pod "30bec9e1-3c58-400b-86fe-0ddcdde2bf8b" (UID: "30bec9e1-3c58-400b-86fe-0ddcdde2bf8b"). InnerVolumeSpecName "kube-api-access-w855q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.161121 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-config-data" (OuterVolumeSpecName: "config-data") pod "30bec9e1-3c58-400b-86fe-0ddcdde2bf8b" (UID: "30bec9e1-3c58-400b-86fe-0ddcdde2bf8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.183850 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f985748e-24cd-4d1b-abe8-7a579aee87af" (UID: "f985748e-24cd-4d1b-abe8-7a579aee87af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.198827 4857 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.198851 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6tbd\" (UniqueName: \"kubernetes.io/projected/f985748e-24cd-4d1b-abe8-7a579aee87af-kube-api-access-d6tbd\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.198864 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.198874 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.198882 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.198891 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w855q\" (UniqueName: \"kubernetes.io/projected/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-kube-api-access-w855q\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.198899 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.198907 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.215553 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-config-data" (OuterVolumeSpecName: "config-data") pod "f985748e-24cd-4d1b-abe8-7a579aee87af" (UID: "f985748e-24cd-4d1b-abe8-7a579aee87af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.278922 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.301746 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f985748e-24cd-4d1b-abe8-7a579aee87af-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.402950 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dg7z6\" (UniqueName: \"kubernetes.io/projected/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-kube-api-access-dg7z6\") pod \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.403105 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-fernet-keys\") pod \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.403181 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-config-data\") pod \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.403301 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-combined-ca-bundle\") pod \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.403406 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-credential-keys\") pod \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.403532 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-scripts\") pod \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\" (UID: \"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e\") " Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.409018 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "54a6f0ca-e3e9-47df-bca6-3e4287d5c32e" (UID: "54a6f0ca-e3e9-47df-bca6-3e4287d5c32e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.409093 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-scripts" (OuterVolumeSpecName: "scripts") pod "54a6f0ca-e3e9-47df-bca6-3e4287d5c32e" (UID: "54a6f0ca-e3e9-47df-bca6-3e4287d5c32e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.409208 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-kube-api-access-dg7z6" (OuterVolumeSpecName: "kube-api-access-dg7z6") pod "54a6f0ca-e3e9-47df-bca6-3e4287d5c32e" (UID: "54a6f0ca-e3e9-47df-bca6-3e4287d5c32e"). InnerVolumeSpecName "kube-api-access-dg7z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.409868 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "54a6f0ca-e3e9-47df-bca6-3e4287d5c32e" (UID: "54a6f0ca-e3e9-47df-bca6-3e4287d5c32e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.432556 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-config-data" (OuterVolumeSpecName: "config-data") pod "54a6f0ca-e3e9-47df-bca6-3e4287d5c32e" (UID: "54a6f0ca-e3e9-47df-bca6-3e4287d5c32e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.452486 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54a6f0ca-e3e9-47df-bca6-3e4287d5c32e" (UID: "54a6f0ca-e3e9-47df-bca6-3e4287d5c32e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.505673 4857 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.505708 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.505719 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.505731 4857 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.505739 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.505747 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dg7z6\" (UniqueName: \"kubernetes.io/projected/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e-kube-api-access-dg7z6\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.930117 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f166659f-d3ef-4f73-8258-46f19eacd6d2","Type":"ContainerStarted","Data":"7936e0e57d464af88b39cfa9e719b2e64386ed51d93bf91ea00cba4b6f4e14b9"} Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.933100 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nq6w4" event={"ID":"54a6f0ca-e3e9-47df-bca6-3e4287d5c32e","Type":"ContainerDied","Data":"3130f0155ff3566e47ed83d1be6720ca8282f869599701034d163346b01391fd"} Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.933149 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3130f0155ff3566e47ed83d1be6720ca8282f869599701034d163346b01391fd" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.933210 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pkbkh" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.933318 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9f69d" Dec 01 21:55:16 crc kubenswrapper[4857]: I1201 21:55:16.938339 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nq6w4" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.050581 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-fdd4494c7-87llf"] Dec 01 21:55:17 crc kubenswrapper[4857]: E1201 21:55:17.050980 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f985748e-24cd-4d1b-abe8-7a579aee87af" containerName="glance-db-sync" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.050996 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f985748e-24cd-4d1b-abe8-7a579aee87af" containerName="glance-db-sync" Dec 01 21:55:17 crc kubenswrapper[4857]: E1201 21:55:17.051010 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30bec9e1-3c58-400b-86fe-0ddcdde2bf8b" containerName="placement-db-sync" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.051017 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="30bec9e1-3c58-400b-86fe-0ddcdde2bf8b" containerName="placement-db-sync" Dec 01 21:55:17 crc kubenswrapper[4857]: E1201 21:55:17.051032 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a6f0ca-e3e9-47df-bca6-3e4287d5c32e" containerName="keystone-bootstrap" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.051073 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a6f0ca-e3e9-47df-bca6-3e4287d5c32e" containerName="keystone-bootstrap" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.051263 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f985748e-24cd-4d1b-abe8-7a579aee87af" containerName="glance-db-sync" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.051281 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="30bec9e1-3c58-400b-86fe-0ddcdde2bf8b" containerName="placement-db-sync" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.051295 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="54a6f0ca-e3e9-47df-bca6-3e4287d5c32e" containerName="keystone-bootstrap" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.051864 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.054884 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.055141 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.055147 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.055373 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6ll6v" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.055662 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.056014 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.087385 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fdd4494c7-87llf"] Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.187910 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6965455568-h4s4z"] Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.190337 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.192962 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.193757 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.193784 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.194005 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-dd4l5" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.194269 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.214137 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6965455568-h4s4z"] Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.222473 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlxf7\" (UniqueName: \"kubernetes.io/projected/085e66d4-f84a-47e2-851b-99f08ccc9683-kube-api-access-dlxf7\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.222690 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-fernet-keys\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.222812 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-combined-ca-bundle\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.222909 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-credential-keys\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.223063 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-internal-tls-certs\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.223164 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-public-tls-certs\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.223267 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-config-data\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.223389 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-scripts\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325326 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-config-data\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325670 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6twm\" (UniqueName: \"kubernetes.io/projected/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-kube-api-access-j6twm\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325710 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-config-data\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325731 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-internal-tls-certs\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325755 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-logs\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325782 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-combined-ca-bundle\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325799 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-scripts\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325818 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlxf7\" (UniqueName: \"kubernetes.io/projected/085e66d4-f84a-47e2-851b-99f08ccc9683-kube-api-access-dlxf7\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325832 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-fernet-keys\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325864 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-scripts\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325883 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-combined-ca-bundle\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325912 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-credential-keys\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325943 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-internal-tls-certs\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325965 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-public-tls-certs\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.325996 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-public-tls-certs\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.331526 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-config-data\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.338583 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-internal-tls-certs\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.350765 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-fernet-keys\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.355562 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-credential-keys\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.355757 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-scripts\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.360680 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-public-tls-certs\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.364976 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085e66d4-f84a-47e2-851b-99f08ccc9683-combined-ca-bundle\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.365309 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlxf7\" (UniqueName: \"kubernetes.io/projected/085e66d4-f84a-47e2-851b-99f08ccc9683-kube-api-access-dlxf7\") pod \"keystone-fdd4494c7-87llf\" (UID: \"085e66d4-f84a-47e2-851b-99f08ccc9683\") " pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.371618 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.429031 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-logs\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.429133 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-combined-ca-bundle\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.429197 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-scripts\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.429292 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-public-tls-certs\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.429364 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6twm\" (UniqueName: \"kubernetes.io/projected/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-kube-api-access-j6twm\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.429390 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-config-data\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.429418 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-internal-tls-certs\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.430907 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-logs\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.434729 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-scripts\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.451387 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-internal-tls-certs\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.452223 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-combined-ca-bundle\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.460189 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-k2mk8"] Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.465115 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-config-data\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.466981 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-public-tls-certs\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.467181 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.473233 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6twm\" (UniqueName: \"kubernetes.io/projected/338abaa2-77aa-403c-b26b-6a5d8c02ff9f-kube-api-access-j6twm\") pod \"placement-6965455568-h4s4z\" (UID: \"338abaa2-77aa-403c-b26b-6a5d8c02ff9f\") " pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.478618 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-k2mk8"] Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.557810 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.638852 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.638946 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qpgg\" (UniqueName: \"kubernetes.io/projected/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-kube-api-access-4qpgg\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.638986 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.639014 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.639068 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-config\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.639090 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.740226 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.741590 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qpgg\" (UniqueName: \"kubernetes.io/projected/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-kube-api-access-4qpgg\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.741615 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.741642 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.741678 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-config\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.741699 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.741341 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.743184 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.743356 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.743983 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-config\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.744489 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.767099 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qpgg\" (UniqueName: \"kubernetes.io/projected/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-kube-api-access-4qpgg\") pod \"dnsmasq-dns-785d8bcb8c-k2mk8\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:17 crc kubenswrapper[4857]: I1201 21:55:17.948158 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.122641 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fdd4494c7-87llf"] Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.402024 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6965455568-h4s4z"] Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.449571 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.451797 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.453888 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bjgmp" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.455370 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.455398 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.456496 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:55:18 crc kubenswrapper[4857]: W1201 21:55:18.474283 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod338abaa2_77aa_403c_b26b_6a5d8c02ff9f.slice/crio-6405a04d9a889adbf6adaa24415f958dd40135c61e00152048d3fefcb448d015 WatchSource:0}: Error finding container 6405a04d9a889adbf6adaa24415f958dd40135c61e00152048d3fefcb448d015: Status 404 returned error can't find the container with id 6405a04d9a889adbf6adaa24415f958dd40135c61e00152048d3fefcb448d015 Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.567729 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-scripts\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.567804 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7nxf\" (UniqueName: \"kubernetes.io/projected/d27f5397-78ae-4289-a153-343f8ed889fa-kube-api-access-s7nxf\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.567847 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d27f5397-78ae-4289-a153-343f8ed889fa-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.567908 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-config-data\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.567930 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d27f5397-78ae-4289-a153-343f8ed889fa-logs\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.567963 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.568032 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.616391 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-k2mk8"] Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.669354 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-scripts\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.669413 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7nxf\" (UniqueName: \"kubernetes.io/projected/d27f5397-78ae-4289-a153-343f8ed889fa-kube-api-access-s7nxf\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.669458 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d27f5397-78ae-4289-a153-343f8ed889fa-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.669506 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-config-data\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.669522 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d27f5397-78ae-4289-a153-343f8ed889fa-logs\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.669546 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.669598 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.670242 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d27f5397-78ae-4289-a153-343f8ed889fa-logs\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.670436 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d27f5397-78ae-4289-a153-343f8ed889fa-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.671603 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.678289 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-config-data\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.678889 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-scripts\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.679856 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.692292 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7nxf\" (UniqueName: \"kubernetes.io/projected/d27f5397-78ae-4289-a153-343f8ed889fa-kube-api-access-s7nxf\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.726091 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.786822 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.788629 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.792509 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.798772 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.806796 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.880726 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.880805 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.880823 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.880845 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a96d656-f8d9-4fc3-8f0e-873f16749256-logs\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.880920 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwlcq\" (UniqueName: \"kubernetes.io/projected/4a96d656-f8d9-4fc3-8f0e-873f16749256-kube-api-access-lwlcq\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.880954 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a96d656-f8d9-4fc3-8f0e-873f16749256-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.881010 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.986187 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.986661 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.986683 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.986704 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a96d656-f8d9-4fc3-8f0e-873f16749256-logs\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.986744 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwlcq\" (UniqueName: \"kubernetes.io/projected/4a96d656-f8d9-4fc3-8f0e-873f16749256-kube-api-access-lwlcq\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.986767 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a96d656-f8d9-4fc3-8f0e-873f16749256-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.986834 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.987581 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a96d656-f8d9-4fc3-8f0e-873f16749256-logs\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.987626 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a96d656-f8d9-4fc3-8f0e-873f16749256-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:18 crc kubenswrapper[4857]: I1201 21:55:18.988093 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.004419 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.009854 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.010491 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.010932 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwlcq\" (UniqueName: \"kubernetes.io/projected/4a96d656-f8d9-4fc3-8f0e-873f16749256-kube-api-access-lwlcq\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.013289 4857 generic.go:334] "Generic (PLEG): container finished" podID="8dc6c548-e991-444e-a627-67896bf3e307" containerID="abf1b62202614efcf7bfea038b20ef4221705a3afe75a406ca46b3134cca079a" exitCode=0 Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.013355 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-s58fk" event={"ID":"8dc6c548-e991-444e-a627-67896bf3e307","Type":"ContainerDied","Data":"abf1b62202614efcf7bfea038b20ef4221705a3afe75a406ca46b3134cca079a"} Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.024303 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6965455568-h4s4z" event={"ID":"338abaa2-77aa-403c-b26b-6a5d8c02ff9f","Type":"ContainerStarted","Data":"3e52fb1c6847a05d8ebbcb4e685b1bc9019f12dcaee9eb5e29f438d41b5e61bc"} Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.024349 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6965455568-h4s4z" event={"ID":"338abaa2-77aa-403c-b26b-6a5d8c02ff9f","Type":"ContainerStarted","Data":"6405a04d9a889adbf6adaa24415f958dd40135c61e00152048d3fefcb448d015"} Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.027119 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fdd4494c7-87llf" event={"ID":"085e66d4-f84a-47e2-851b-99f08ccc9683","Type":"ContainerStarted","Data":"85afbaa112c4314bfa5b1b21b6b31b69f3f4f2dc3cfbd9631728e1c77208aec4"} Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.027145 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fdd4494c7-87llf" event={"ID":"085e66d4-f84a-47e2-851b-99f08ccc9683","Type":"ContainerStarted","Data":"e0ac4ba1b5a4cf53cc2591e0e8b2ce4c1bb18c1d8eab60be931e90b85beeb995"} Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.028061 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.042669 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" event={"ID":"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a","Type":"ContainerStarted","Data":"ac517e33864dfb3ecdf2639c58b35a7c6519f3fc2848d4c40d74312ea6bc0867"} Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.042708 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" event={"ID":"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a","Type":"ContainerStarted","Data":"db99bbebffeb43f69523599cdba1ee87f0d13c0b318c14bb2915af773dc37255"} Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.051117 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.062008 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-fdd4494c7-87llf" podStartSLOduration=2.061993843 podStartE2EDuration="2.061993843s" podCreationTimestamp="2025-12-01 21:55:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:19.05535051 +0000 UTC m=+1237.545412827" watchObservedRunningTime="2025-12-01 21:55:19.061993843 +0000 UTC m=+1237.552056160" Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.113099 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.691224 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:55:19 crc kubenswrapper[4857]: W1201 21:55:19.726234 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd27f5397_78ae_4289_a153_343f8ed889fa.slice/crio-9d1b17a1e6cb4f6755136ca47182056cb7db92fbacd866a5025db7ca7969777d WatchSource:0}: Error finding container 9d1b17a1e6cb4f6755136ca47182056cb7db92fbacd866a5025db7ca7969777d: Status 404 returned error can't find the container with id 9d1b17a1e6cb4f6755136ca47182056cb7db92fbacd866a5025db7ca7969777d Dec 01 21:55:19 crc kubenswrapper[4857]: I1201 21:55:19.972149 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:55:20 crc kubenswrapper[4857]: W1201 21:55:20.003003 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a96d656_f8d9_4fc3_8f0e_873f16749256.slice/crio-4a5ebe1e0a4b79579b1119db303b809b615f1602e01803872279d70b02dfbf67 WatchSource:0}: Error finding container 4a5ebe1e0a4b79579b1119db303b809b615f1602e01803872279d70b02dfbf67: Status 404 returned error can't find the container with id 4a5ebe1e0a4b79579b1119db303b809b615f1602e01803872279d70b02dfbf67 Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.079014 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6965455568-h4s4z" event={"ID":"338abaa2-77aa-403c-b26b-6a5d8c02ff9f","Type":"ContainerStarted","Data":"160fe6431c7e9d81ec71698fb9ce63d9c18b054a37aaff12a90d12dfe30f196e"} Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.080553 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.080592 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.118434 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4a96d656-f8d9-4fc3-8f0e-873f16749256","Type":"ContainerStarted","Data":"4a5ebe1e0a4b79579b1119db303b809b615f1602e01803872279d70b02dfbf67"} Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.128884 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6965455568-h4s4z" podStartSLOduration=3.128860224 podStartE2EDuration="3.128860224s" podCreationTimestamp="2025-12-01 21:55:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:20.108989356 +0000 UTC m=+1238.599051673" watchObservedRunningTime="2025-12-01 21:55:20.128860224 +0000 UTC m=+1238.618922541" Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.135570 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d27f5397-78ae-4289-a153-343f8ed889fa","Type":"ContainerStarted","Data":"9d1b17a1e6cb4f6755136ca47182056cb7db92fbacd866a5025db7ca7969777d"} Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.144321 4857 generic.go:334] "Generic (PLEG): container finished" podID="b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" containerID="ac517e33864dfb3ecdf2639c58b35a7c6519f3fc2848d4c40d74312ea6bc0867" exitCode=0 Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.144380 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" event={"ID":"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a","Type":"ContainerDied","Data":"ac517e33864dfb3ecdf2639c58b35a7c6519f3fc2848d4c40d74312ea6bc0867"} Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.144447 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" event={"ID":"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a","Type":"ContainerStarted","Data":"80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb"} Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.188682 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" podStartSLOduration=3.188645579 podStartE2EDuration="3.188645579s" podCreationTimestamp="2025-12-01 21:55:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:20.17889228 +0000 UTC m=+1238.668954617" watchObservedRunningTime="2025-12-01 21:55:20.188645579 +0000 UTC m=+1238.678707896" Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.596414 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-s58fk" Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.632784 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dc6c548-e991-444e-a627-67896bf3e307-combined-ca-bundle\") pod \"8dc6c548-e991-444e-a627-67896bf3e307\" (UID: \"8dc6c548-e991-444e-a627-67896bf3e307\") " Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.632886 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8dc6c548-e991-444e-a627-67896bf3e307-config\") pod \"8dc6c548-e991-444e-a627-67896bf3e307\" (UID: \"8dc6c548-e991-444e-a627-67896bf3e307\") " Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.633030 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ht2s\" (UniqueName: \"kubernetes.io/projected/8dc6c548-e991-444e-a627-67896bf3e307-kube-api-access-9ht2s\") pod \"8dc6c548-e991-444e-a627-67896bf3e307\" (UID: \"8dc6c548-e991-444e-a627-67896bf3e307\") " Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.651228 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dc6c548-e991-444e-a627-67896bf3e307-kube-api-access-9ht2s" (OuterVolumeSpecName: "kube-api-access-9ht2s") pod "8dc6c548-e991-444e-a627-67896bf3e307" (UID: "8dc6c548-e991-444e-a627-67896bf3e307"). InnerVolumeSpecName "kube-api-access-9ht2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.679742 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dc6c548-e991-444e-a627-67896bf3e307-config" (OuterVolumeSpecName: "config") pod "8dc6c548-e991-444e-a627-67896bf3e307" (UID: "8dc6c548-e991-444e-a627-67896bf3e307"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.724118 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dc6c548-e991-444e-a627-67896bf3e307-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8dc6c548-e991-444e-a627-67896bf3e307" (UID: "8dc6c548-e991-444e-a627-67896bf3e307"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.735427 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ht2s\" (UniqueName: \"kubernetes.io/projected/8dc6c548-e991-444e-a627-67896bf3e307-kube-api-access-9ht2s\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.735481 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dc6c548-e991-444e-a627-67896bf3e307-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.735495 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8dc6c548-e991-444e-a627-67896bf3e307-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.811202 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:55:20 crc kubenswrapper[4857]: I1201 21:55:20.863126 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.162983 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-s58fk" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.164297 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-s58fk" event={"ID":"8dc6c548-e991-444e-a627-67896bf3e307","Type":"ContainerDied","Data":"41fbc7226100123d3f5439293d825e540538c03d4376f293d795de38e0b1d2ac"} Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.164337 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41fbc7226100123d3f5439293d825e540538c03d4376f293d795de38e0b1d2ac" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.164756 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.233641 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-k2mk8"] Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.290879 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-mvhhh"] Dec 01 21:55:21 crc kubenswrapper[4857]: E1201 21:55:21.291480 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc6c548-e991-444e-a627-67896bf3e307" containerName="neutron-db-sync" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.291505 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc6c548-e991-444e-a627-67896bf3e307" containerName="neutron-db-sync" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.291799 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dc6c548-e991-444e-a627-67896bf3e307" containerName="neutron-db-sync" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.297564 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.308878 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-mvhhh"] Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.351483 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.351553 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.351607 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.351655 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-config\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.351705 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2dfn\" (UniqueName: \"kubernetes.io/projected/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-kube-api-access-f2dfn\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.351812 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.434032 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7cf98975f6-8bj6t"] Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.437709 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.443192 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.443380 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-6jgxw" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.443523 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.443980 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.454980 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-config\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.455061 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q69kp\" (UniqueName: \"kubernetes.io/projected/e04bffff-693d-4909-914d-1e2f784ca47d-kube-api-access-q69kp\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.455149 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-httpd-config\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.455281 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2dfn\" (UniqueName: \"kubernetes.io/projected/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-kube-api-access-f2dfn\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.455393 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-config\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.455436 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.455541 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.455619 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.455719 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.455754 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-combined-ca-bundle\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.455813 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-ovndb-tls-certs\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.455884 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-config\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.456972 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.457637 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.457821 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.458001 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.461388 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7cf98975f6-8bj6t"] Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.494790 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2dfn\" (UniqueName: \"kubernetes.io/projected/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-kube-api-access-f2dfn\") pod \"dnsmasq-dns-55f844cf75-mvhhh\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.557735 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-combined-ca-bundle\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.557829 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-ovndb-tls-certs\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.557892 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q69kp\" (UniqueName: \"kubernetes.io/projected/e04bffff-693d-4909-914d-1e2f784ca47d-kube-api-access-q69kp\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.557962 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-httpd-config\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.558036 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-config\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.570937 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-combined-ca-bundle\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.572589 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-ovndb-tls-certs\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.576188 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q69kp\" (UniqueName: \"kubernetes.io/projected/e04bffff-693d-4909-914d-1e2f784ca47d-kube-api-access-q69kp\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.583800 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-httpd-config\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.586532 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-config\") pod \"neutron-7cf98975f6-8bj6t\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.622579 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:21 crc kubenswrapper[4857]: I1201 21:55:21.761957 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:22 crc kubenswrapper[4857]: I1201 21:55:22.226057 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-mvhhh"] Dec 01 21:55:22 crc kubenswrapper[4857]: I1201 21:55:22.521752 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7cf98975f6-8bj6t"] Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.246168 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cf98975f6-8bj6t" event={"ID":"e04bffff-693d-4909-914d-1e2f784ca47d","Type":"ContainerStarted","Data":"c22bf24c46ff5fd423c0066118d5059901c95219f85fd49109d8fe88fe75264d"} Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.246863 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cf98975f6-8bj6t" event={"ID":"e04bffff-693d-4909-914d-1e2f784ca47d","Type":"ContainerStarted","Data":"87669aaec6ee97132b534dc4cb15df194c0ffa37824141d4493d15dd3792e69e"} Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.251915 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4a96d656-f8d9-4fc3-8f0e-873f16749256","Type":"ContainerStarted","Data":"448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42"} Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.255234 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d27f5397-78ae-4289-a153-343f8ed889fa","Type":"ContainerStarted","Data":"d665290d31a51681d8f8c4e4b0751cb4fb073752e3e30f19fc9b4381f051f96e"} Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.257886 4857 generic.go:334] "Generic (PLEG): container finished" podID="6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" containerID="325d96063eaaed29781d6ebcbce77f4b3fb0988161519e923c1d12aaad24f6e3" exitCode=0 Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.258273 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" podUID="b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" containerName="dnsmasq-dns" containerID="cri-o://80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb" gracePeriod=10 Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.259316 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" event={"ID":"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a","Type":"ContainerDied","Data":"325d96063eaaed29781d6ebcbce77f4b3fb0988161519e923c1d12aaad24f6e3"} Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.262656 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" event={"ID":"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a","Type":"ContainerStarted","Data":"d0232fd272fcc87c94d9c72349d92cbb28bdbed8ef25a02502d3612f379e25b5"} Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.798280 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.948428 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-dns-swift-storage-0\") pod \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.948608 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qpgg\" (UniqueName: \"kubernetes.io/projected/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-kube-api-access-4qpgg\") pod \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.948672 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-config\") pod \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.948751 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-dns-svc\") pod \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.948812 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-ovsdbserver-sb\") pod \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.948861 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-ovsdbserver-nb\") pod \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\" (UID: \"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a\") " Dec 01 21:55:23 crc kubenswrapper[4857]: I1201 21:55:23.967698 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-kube-api-access-4qpgg" (OuterVolumeSpecName: "kube-api-access-4qpgg") pod "b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" (UID: "b0e9f945-25e6-42a1-aa98-ed445bdcdc3a"). InnerVolumeSpecName "kube-api-access-4qpgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.033922 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" (UID: "b0e9f945-25e6-42a1-aa98-ed445bdcdc3a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.038274 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-config" (OuterVolumeSpecName: "config") pod "b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" (UID: "b0e9f945-25e6-42a1-aa98-ed445bdcdc3a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.047971 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" (UID: "b0e9f945-25e6-42a1-aa98-ed445bdcdc3a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.058048 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.058206 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.058271 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qpgg\" (UniqueName: \"kubernetes.io/projected/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-kube-api-access-4qpgg\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.058337 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.079357 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" (UID: "b0e9f945-25e6-42a1-aa98-ed445bdcdc3a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.080027 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" (UID: "b0e9f945-25e6-42a1-aa98-ed445bdcdc3a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.173726 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.174173 4857 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.270956 4857 generic.go:334] "Generic (PLEG): container finished" podID="b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" containerID="80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb" exitCode=0 Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.271060 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" event={"ID":"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a","Type":"ContainerDied","Data":"80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb"} Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.271101 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" event={"ID":"b0e9f945-25e6-42a1-aa98-ed445bdcdc3a","Type":"ContainerDied","Data":"db99bbebffeb43f69523599cdba1ee87f0d13c0b318c14bb2915af773dc37255"} Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.271124 4857 scope.go:117] "RemoveContainer" containerID="80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.271241 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-k2mk8" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.276067 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cf98975f6-8bj6t" event={"ID":"e04bffff-693d-4909-914d-1e2f784ca47d","Type":"ContainerStarted","Data":"95e8a45ce13de4d4d028b19ce5c61af5facfe7703e3996a78f3f2f220ccc1303"} Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.277887 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.293748 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ppkc6" event={"ID":"66c9aa39-8cad-4807-98c7-ee329d87bb20","Type":"ContainerStarted","Data":"a7330b1386f3883c5a5a6f43655ee634bc0e7398139f79dd64fc1c3c79b7bd42"} Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.296954 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xzd26" event={"ID":"aa7ace20-c228-447f-98b7-dc09e562d5b4","Type":"ContainerStarted","Data":"7d379a633511ffa997abc9a5932efaa8f962b1162a5507e79fe062fcd85a4ce1"} Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.307839 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4a96d656-f8d9-4fc3-8f0e-873f16749256","Type":"ContainerStarted","Data":"eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3"} Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.307886 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4a96d656-f8d9-4fc3-8f0e-873f16749256" containerName="glance-log" containerID="cri-o://448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42" gracePeriod=30 Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.307980 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4a96d656-f8d9-4fc3-8f0e-873f16749256" containerName="glance-httpd" containerID="cri-o://eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3" gracePeriod=30 Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.313158 4857 scope.go:117] "RemoveContainer" containerID="ac517e33864dfb3ecdf2639c58b35a7c6519f3fc2848d4c40d74312ea6bc0867" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.331658 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d27f5397-78ae-4289-a153-343f8ed889fa","Type":"ContainerStarted","Data":"9df5f41182fb925d21f9486bde5ef4d9bf7bb43ce5b8371d3334430802a937b4"} Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.333389 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d27f5397-78ae-4289-a153-343f8ed889fa" containerName="glance-log" containerID="cri-o://d665290d31a51681d8f8c4e4b0751cb4fb073752e3e30f19fc9b4381f051f96e" gracePeriod=30 Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.333521 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d27f5397-78ae-4289-a153-343f8ed889fa" containerName="glance-httpd" containerID="cri-o://9df5f41182fb925d21f9486bde5ef4d9bf7bb43ce5b8371d3334430802a937b4" gracePeriod=30 Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.337649 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7cf98975f6-8bj6t" podStartSLOduration=3.337637327 podStartE2EDuration="3.337637327s" podCreationTimestamp="2025-12-01 21:55:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:24.326253038 +0000 UTC m=+1242.816315375" watchObservedRunningTime="2025-12-01 21:55:24.337637327 +0000 UTC m=+1242.827699644" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.341309 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" event={"ID":"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a","Type":"ContainerStarted","Data":"4ee1494f162df42894e78241882217b883d847b8735ae5641691eb6d19e7548a"} Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.342382 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.350526 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-57b8cf5fc7-kg4kw"] Dec 01 21:55:24 crc kubenswrapper[4857]: E1201 21:55:24.352566 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" containerName="dnsmasq-dns" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.352589 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" containerName="dnsmasq-dns" Dec 01 21:55:24 crc kubenswrapper[4857]: E1201 21:55:24.352628 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" containerName="init" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.352636 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" containerName="init" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.352847 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" containerName="dnsmasq-dns" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.353748 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.360003 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.360275 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.363920 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-57b8cf5fc7-kg4kw"] Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.364747 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.364725101 podStartE2EDuration="7.364725101s" podCreationTimestamp="2025-12-01 21:55:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:24.351677472 +0000 UTC m=+1242.841739789" watchObservedRunningTime="2025-12-01 21:55:24.364725101 +0000 UTC m=+1242.854787418" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.376470 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-ppkc6" podStartSLOduration=3.707672483 podStartE2EDuration="54.376449709s" podCreationTimestamp="2025-12-01 21:54:30 +0000 UTC" firstStartedPulling="2025-12-01 21:54:32.22746115 +0000 UTC m=+1190.717523467" lastFinishedPulling="2025-12-01 21:55:22.896238376 +0000 UTC m=+1241.386300693" observedRunningTime="2025-12-01 21:55:24.372114833 +0000 UTC m=+1242.862177150" watchObservedRunningTime="2025-12-01 21:55:24.376449709 +0000 UTC m=+1242.866512026" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.390163 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bws8d\" (UniqueName: \"kubernetes.io/projected/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-kube-api-access-bws8d\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.392836 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-public-tls-certs\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.392968 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-ovndb-tls-certs\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.393246 4857 scope.go:117] "RemoveContainer" containerID="80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.394138 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-config\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.394710 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-internal-tls-certs\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.394904 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-combined-ca-bundle\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.395011 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-httpd-config\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: E1201 21:55:24.395293 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb\": container with ID starting with 80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb not found: ID does not exist" containerID="80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.395339 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb"} err="failed to get container status \"80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb\": rpc error: code = NotFound desc = could not find container \"80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb\": container with ID starting with 80f2cf45cf844de2c01837562e6b82966436a111590cb9242115aa0a5f8bcfdb not found: ID does not exist" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.395376 4857 scope.go:117] "RemoveContainer" containerID="ac517e33864dfb3ecdf2639c58b35a7c6519f3fc2848d4c40d74312ea6bc0867" Dec 01 21:55:24 crc kubenswrapper[4857]: E1201 21:55:24.402018 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac517e33864dfb3ecdf2639c58b35a7c6519f3fc2848d4c40d74312ea6bc0867\": container with ID starting with ac517e33864dfb3ecdf2639c58b35a7c6519f3fc2848d4c40d74312ea6bc0867 not found: ID does not exist" containerID="ac517e33864dfb3ecdf2639c58b35a7c6519f3fc2848d4c40d74312ea6bc0867" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.402089 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac517e33864dfb3ecdf2639c58b35a7c6519f3fc2848d4c40d74312ea6bc0867"} err="failed to get container status \"ac517e33864dfb3ecdf2639c58b35a7c6519f3fc2848d4c40d74312ea6bc0867\": rpc error: code = NotFound desc = could not find container \"ac517e33864dfb3ecdf2639c58b35a7c6519f3fc2848d4c40d74312ea6bc0867\": container with ID starting with ac517e33864dfb3ecdf2639c58b35a7c6519f3fc2848d4c40d74312ea6bc0867 not found: ID does not exist" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.408778 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-xzd26" podStartSLOduration=3.241319092 podStartE2EDuration="54.408752421s" podCreationTimestamp="2025-12-01 21:54:30 +0000 UTC" firstStartedPulling="2025-12-01 21:54:31.922461464 +0000 UTC m=+1190.412523781" lastFinishedPulling="2025-12-01 21:55:23.089894773 +0000 UTC m=+1241.579957110" observedRunningTime="2025-12-01 21:55:24.402448336 +0000 UTC m=+1242.892510653" watchObservedRunningTime="2025-12-01 21:55:24.408752421 +0000 UTC m=+1242.898814738" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.437268 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-k2mk8"] Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.512923 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-public-tls-certs\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.516556 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-ovndb-tls-certs\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.516687 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-config\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.517200 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-internal-tls-certs\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.517376 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-combined-ca-bundle\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.517431 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-httpd-config\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.517571 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bws8d\" (UniqueName: \"kubernetes.io/projected/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-kube-api-access-bws8d\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.520528 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-k2mk8"] Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.524524 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-ovndb-tls-certs\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.528952 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-internal-tls-certs\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.529260 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-config\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.529404 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-public-tls-certs\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.530818 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-combined-ca-bundle\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.533026 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-httpd-config\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.541905 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.541874384 podStartE2EDuration="7.541874384s" podCreationTimestamp="2025-12-01 21:55:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:24.457744591 +0000 UTC m=+1242.947806908" watchObservedRunningTime="2025-12-01 21:55:24.541874384 +0000 UTC m=+1243.031936701" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.543873 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" podStartSLOduration=3.543864472 podStartE2EDuration="3.543864472s" podCreationTimestamp="2025-12-01 21:55:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:24.486986378 +0000 UTC m=+1242.977048695" watchObservedRunningTime="2025-12-01 21:55:24.543864472 +0000 UTC m=+1243.033926779" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.544943 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bws8d\" (UniqueName: \"kubernetes.io/projected/33d2facf-f0c8-4566-a15c-038f0f9f9a6e-kube-api-access-bws8d\") pod \"neutron-57b8cf5fc7-kg4kw\" (UID: \"33d2facf-f0c8-4566-a15c-038f0f9f9a6e\") " pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:24 crc kubenswrapper[4857]: I1201 21:55:24.594650 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.108518 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.244653 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a96d656-f8d9-4fc3-8f0e-873f16749256-logs\") pod \"4a96d656-f8d9-4fc3-8f0e-873f16749256\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.244721 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a96d656-f8d9-4fc3-8f0e-873f16749256-httpd-run\") pod \"4a96d656-f8d9-4fc3-8f0e-873f16749256\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.244748 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-combined-ca-bundle\") pod \"4a96d656-f8d9-4fc3-8f0e-873f16749256\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.244903 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-config-data\") pod \"4a96d656-f8d9-4fc3-8f0e-873f16749256\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.244951 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-scripts\") pod \"4a96d656-f8d9-4fc3-8f0e-873f16749256\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.245017 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwlcq\" (UniqueName: \"kubernetes.io/projected/4a96d656-f8d9-4fc3-8f0e-873f16749256-kube-api-access-lwlcq\") pod \"4a96d656-f8d9-4fc3-8f0e-873f16749256\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.245114 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"4a96d656-f8d9-4fc3-8f0e-873f16749256\" (UID: \"4a96d656-f8d9-4fc3-8f0e-873f16749256\") " Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.253402 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a96d656-f8d9-4fc3-8f0e-873f16749256-logs" (OuterVolumeSpecName: "logs") pod "4a96d656-f8d9-4fc3-8f0e-873f16749256" (UID: "4a96d656-f8d9-4fc3-8f0e-873f16749256"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.253636 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a96d656-f8d9-4fc3-8f0e-873f16749256-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4a96d656-f8d9-4fc3-8f0e-873f16749256" (UID: "4a96d656-f8d9-4fc3-8f0e-873f16749256"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.277321 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a96d656-f8d9-4fc3-8f0e-873f16749256-kube-api-access-lwlcq" (OuterVolumeSpecName: "kube-api-access-lwlcq") pod "4a96d656-f8d9-4fc3-8f0e-873f16749256" (UID: "4a96d656-f8d9-4fc3-8f0e-873f16749256"). InnerVolumeSpecName "kube-api-access-lwlcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.280271 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-scripts" (OuterVolumeSpecName: "scripts") pod "4a96d656-f8d9-4fc3-8f0e-873f16749256" (UID: "4a96d656-f8d9-4fc3-8f0e-873f16749256"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.292231 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "4a96d656-f8d9-4fc3-8f0e-873f16749256" (UID: "4a96d656-f8d9-4fc3-8f0e-873f16749256"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.350827 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.350861 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwlcq\" (UniqueName: \"kubernetes.io/projected/4a96d656-f8d9-4fc3-8f0e-873f16749256-kube-api-access-lwlcq\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.350887 4857 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.350898 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a96d656-f8d9-4fc3-8f0e-873f16749256-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.350908 4857 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a96d656-f8d9-4fc3-8f0e-873f16749256-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.382478 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a96d656-f8d9-4fc3-8f0e-873f16749256" (UID: "4a96d656-f8d9-4fc3-8f0e-873f16749256"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.388895 4857 generic.go:334] "Generic (PLEG): container finished" podID="d27f5397-78ae-4289-a153-343f8ed889fa" containerID="9df5f41182fb925d21f9486bde5ef4d9bf7bb43ce5b8371d3334430802a937b4" exitCode=143 Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.388928 4857 generic.go:334] "Generic (PLEG): container finished" podID="d27f5397-78ae-4289-a153-343f8ed889fa" containerID="d665290d31a51681d8f8c4e4b0751cb4fb073752e3e30f19fc9b4381f051f96e" exitCode=143 Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.388991 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d27f5397-78ae-4289-a153-343f8ed889fa","Type":"ContainerDied","Data":"9df5f41182fb925d21f9486bde5ef4d9bf7bb43ce5b8371d3334430802a937b4"} Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.389020 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d27f5397-78ae-4289-a153-343f8ed889fa","Type":"ContainerDied","Data":"d665290d31a51681d8f8c4e4b0751cb4fb073752e3e30f19fc9b4381f051f96e"} Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.390684 4857 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.412499 4857 generic.go:334] "Generic (PLEG): container finished" podID="4a96d656-f8d9-4fc3-8f0e-873f16749256" containerID="eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3" exitCode=143 Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.412539 4857 generic.go:334] "Generic (PLEG): container finished" podID="4a96d656-f8d9-4fc3-8f0e-873f16749256" containerID="448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42" exitCode=143 Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.413179 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4a96d656-f8d9-4fc3-8f0e-873f16749256","Type":"ContainerDied","Data":"eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3"} Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.413257 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4a96d656-f8d9-4fc3-8f0e-873f16749256","Type":"ContainerDied","Data":"448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42"} Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.413270 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4a96d656-f8d9-4fc3-8f0e-873f16749256","Type":"ContainerDied","Data":"4a5ebe1e0a4b79579b1119db303b809b615f1602e01803872279d70b02dfbf67"} Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.413287 4857 scope.go:117] "RemoveContainer" containerID="eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.413210 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.426832 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-config-data" (OuterVolumeSpecName: "config-data") pod "4a96d656-f8d9-4fc3-8f0e-873f16749256" (UID: "4a96d656-f8d9-4fc3-8f0e-873f16749256"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.451353 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-57b8cf5fc7-kg4kw"] Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.453103 4857 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.453159 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.453179 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a96d656-f8d9-4fc3-8f0e-873f16749256-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.757380 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.777034 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.789794 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:55:25 crc kubenswrapper[4857]: E1201 21:55:25.815218 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a96d656-f8d9-4fc3-8f0e-873f16749256" containerName="glance-httpd" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.815268 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a96d656-f8d9-4fc3-8f0e-873f16749256" containerName="glance-httpd" Dec 01 21:55:25 crc kubenswrapper[4857]: E1201 21:55:25.815339 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a96d656-f8d9-4fc3-8f0e-873f16749256" containerName="glance-log" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.815368 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a96d656-f8d9-4fc3-8f0e-873f16749256" containerName="glance-log" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.815924 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a96d656-f8d9-4fc3-8f0e-873f16749256" containerName="glance-log" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.815952 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a96d656-f8d9-4fc3-8f0e-873f16749256" containerName="glance-httpd" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.817307 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.817431 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.823920 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.824737 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.872055 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a96d656-f8d9-4fc3-8f0e-873f16749256" path="/var/lib/kubelet/pods/4a96d656-f8d9-4fc3-8f0e-873f16749256/volumes" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.877925 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0e9f945-25e6-42a1-aa98-ed445bdcdc3a" path="/var/lib/kubelet/pods/b0e9f945-25e6-42a1-aa98-ed445bdcdc3a/volumes" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.966945 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.967028 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ec1f836-90ab-4f44-b500-5a705a165925-logs\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.967171 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ec1f836-90ab-4f44-b500-5a705a165925-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.967340 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.967391 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.967428 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.967478 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:25 crc kubenswrapper[4857]: I1201 21:55:25.967545 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9fxr\" (UniqueName: \"kubernetes.io/projected/3ec1f836-90ab-4f44-b500-5a705a165925-kube-api-access-m9fxr\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.069912 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ec1f836-90ab-4f44-b500-5a705a165925-logs\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.069994 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ec1f836-90ab-4f44-b500-5a705a165925-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.070074 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.070103 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.070134 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.070169 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.070210 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9fxr\" (UniqueName: \"kubernetes.io/projected/3ec1f836-90ab-4f44-b500-5a705a165925-kube-api-access-m9fxr\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.070258 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.071660 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ec1f836-90ab-4f44-b500-5a705a165925-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.072016 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ec1f836-90ab-4f44-b500-5a705a165925-logs\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.072568 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.078570 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.080392 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.081889 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.083096 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.098290 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9fxr\" (UniqueName: \"kubernetes.io/projected/3ec1f836-90ab-4f44-b500-5a705a165925-kube-api-access-m9fxr\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.113163 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:55:26 crc kubenswrapper[4857]: I1201 21:55:26.152441 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:27 crc kubenswrapper[4857]: I1201 21:55:27.805840 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:55:27 crc kubenswrapper[4857]: I1201 21:55:27.806322 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:55:28 crc kubenswrapper[4857]: I1201 21:55:28.444475 4857 generic.go:334] "Generic (PLEG): container finished" podID="66c9aa39-8cad-4807-98c7-ee329d87bb20" containerID="a7330b1386f3883c5a5a6f43655ee634bc0e7398139f79dd64fc1c3c79b7bd42" exitCode=0 Dec 01 21:55:28 crc kubenswrapper[4857]: I1201 21:55:28.444967 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ppkc6" event={"ID":"66c9aa39-8cad-4807-98c7-ee329d87bb20","Type":"ContainerDied","Data":"a7330b1386f3883c5a5a6f43655ee634bc0e7398139f79dd64fc1c3c79b7bd42"} Dec 01 21:55:29 crc kubenswrapper[4857]: I1201 21:55:29.458251 4857 scope.go:117] "RemoveContainer" containerID="448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42" Dec 01 21:55:29 crc kubenswrapper[4857]: I1201 21:55:29.459867 4857 generic.go:334] "Generic (PLEG): container finished" podID="aa7ace20-c228-447f-98b7-dc09e562d5b4" containerID="7d379a633511ffa997abc9a5932efaa8f962b1162a5507e79fe062fcd85a4ce1" exitCode=0 Dec 01 21:55:29 crc kubenswrapper[4857]: I1201 21:55:29.459985 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xzd26" event={"ID":"aa7ace20-c228-447f-98b7-dc09e562d5b4","Type":"ContainerDied","Data":"7d379a633511ffa997abc9a5932efaa8f962b1162a5507e79fe062fcd85a4ce1"} Dec 01 21:55:29 crc kubenswrapper[4857]: I1201 21:55:29.481555 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57b8cf5fc7-kg4kw" event={"ID":"33d2facf-f0c8-4566-a15c-038f0f9f9a6e","Type":"ContainerStarted","Data":"1555ffd91a89a2d7d3ecb9a444a9d513658539e2107f94d26295f4c824dca71f"} Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.013259 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.020988 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xzd26" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.028992 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096032 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-combined-ca-bundle\") pod \"d27f5397-78ae-4289-a153-343f8ed889fa\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096097 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"d27f5397-78ae-4289-a153-343f8ed889fa\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096160 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-scripts\") pod \"d27f5397-78ae-4289-a153-343f8ed889fa\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096226 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d27f5397-78ae-4289-a153-343f8ed889fa-logs\") pod \"d27f5397-78ae-4289-a153-343f8ed889fa\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096324 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66c9aa39-8cad-4807-98c7-ee329d87bb20-db-sync-config-data\") pod \"66c9aa39-8cad-4807-98c7-ee329d87bb20\" (UID: \"66c9aa39-8cad-4807-98c7-ee329d87bb20\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096370 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjtrt\" (UniqueName: \"kubernetes.io/projected/66c9aa39-8cad-4807-98c7-ee329d87bb20-kube-api-access-pjtrt\") pod \"66c9aa39-8cad-4807-98c7-ee329d87bb20\" (UID: \"66c9aa39-8cad-4807-98c7-ee329d87bb20\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096447 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7ace20-c228-447f-98b7-dc09e562d5b4-combined-ca-bundle\") pod \"aa7ace20-c228-447f-98b7-dc09e562d5b4\" (UID: \"aa7ace20-c228-447f-98b7-dc09e562d5b4\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096497 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgbf7\" (UniqueName: \"kubernetes.io/projected/aa7ace20-c228-447f-98b7-dc09e562d5b4-kube-api-access-dgbf7\") pod \"aa7ace20-c228-447f-98b7-dc09e562d5b4\" (UID: \"aa7ace20-c228-447f-98b7-dc09e562d5b4\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096529 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c9aa39-8cad-4807-98c7-ee329d87bb20-combined-ca-bundle\") pod \"66c9aa39-8cad-4807-98c7-ee329d87bb20\" (UID: \"66c9aa39-8cad-4807-98c7-ee329d87bb20\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096615 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7ace20-c228-447f-98b7-dc09e562d5b4-config-data\") pod \"aa7ace20-c228-447f-98b7-dc09e562d5b4\" (UID: \"aa7ace20-c228-447f-98b7-dc09e562d5b4\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096668 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7nxf\" (UniqueName: \"kubernetes.io/projected/d27f5397-78ae-4289-a153-343f8ed889fa-kube-api-access-s7nxf\") pod \"d27f5397-78ae-4289-a153-343f8ed889fa\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096709 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-config-data\") pod \"d27f5397-78ae-4289-a153-343f8ed889fa\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.096740 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d27f5397-78ae-4289-a153-343f8ed889fa-httpd-run\") pod \"d27f5397-78ae-4289-a153-343f8ed889fa\" (UID: \"d27f5397-78ae-4289-a153-343f8ed889fa\") " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.097693 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d27f5397-78ae-4289-a153-343f8ed889fa-logs" (OuterVolumeSpecName: "logs") pod "d27f5397-78ae-4289-a153-343f8ed889fa" (UID: "d27f5397-78ae-4289-a153-343f8ed889fa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.097855 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d27f5397-78ae-4289-a153-343f8ed889fa-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d27f5397-78ae-4289-a153-343f8ed889fa" (UID: "d27f5397-78ae-4289-a153-343f8ed889fa"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.106799 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c9aa39-8cad-4807-98c7-ee329d87bb20-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "66c9aa39-8cad-4807-98c7-ee329d87bb20" (UID: "66c9aa39-8cad-4807-98c7-ee329d87bb20"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.108790 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66c9aa39-8cad-4807-98c7-ee329d87bb20-kube-api-access-pjtrt" (OuterVolumeSpecName: "kube-api-access-pjtrt") pod "66c9aa39-8cad-4807-98c7-ee329d87bb20" (UID: "66c9aa39-8cad-4807-98c7-ee329d87bb20"). InnerVolumeSpecName "kube-api-access-pjtrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.110741 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-scripts" (OuterVolumeSpecName: "scripts") pod "d27f5397-78ae-4289-a153-343f8ed889fa" (UID: "d27f5397-78ae-4289-a153-343f8ed889fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.111271 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa7ace20-c228-447f-98b7-dc09e562d5b4-kube-api-access-dgbf7" (OuterVolumeSpecName: "kube-api-access-dgbf7") pod "aa7ace20-c228-447f-98b7-dc09e562d5b4" (UID: "aa7ace20-c228-447f-98b7-dc09e562d5b4"). InnerVolumeSpecName "kube-api-access-dgbf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.112387 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d27f5397-78ae-4289-a153-343f8ed889fa-kube-api-access-s7nxf" (OuterVolumeSpecName: "kube-api-access-s7nxf") pod "d27f5397-78ae-4289-a153-343f8ed889fa" (UID: "d27f5397-78ae-4289-a153-343f8ed889fa"). InnerVolumeSpecName "kube-api-access-s7nxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.125560 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "d27f5397-78ae-4289-a153-343f8ed889fa" (UID: "d27f5397-78ae-4289-a153-343f8ed889fa"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.156902 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c9aa39-8cad-4807-98c7-ee329d87bb20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66c9aa39-8cad-4807-98c7-ee329d87bb20" (UID: "66c9aa39-8cad-4807-98c7-ee329d87bb20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.159498 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d27f5397-78ae-4289-a153-343f8ed889fa" (UID: "d27f5397-78ae-4289-a153-343f8ed889fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.165974 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa7ace20-c228-447f-98b7-dc09e562d5b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa7ace20-c228-447f-98b7-dc09e562d5b4" (UID: "aa7ace20-c228-447f-98b7-dc09e562d5b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.181453 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-config-data" (OuterVolumeSpecName: "config-data") pod "d27f5397-78ae-4289-a153-343f8ed889fa" (UID: "d27f5397-78ae-4289-a153-343f8ed889fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.199907 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7nxf\" (UniqueName: \"kubernetes.io/projected/d27f5397-78ae-4289-a153-343f8ed889fa-kube-api-access-s7nxf\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.199942 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.199953 4857 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d27f5397-78ae-4289-a153-343f8ed889fa-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.199982 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.200016 4857 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.200028 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d27f5397-78ae-4289-a153-343f8ed889fa-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.200068 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d27f5397-78ae-4289-a153-343f8ed889fa-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.200080 4857 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66c9aa39-8cad-4807-98c7-ee329d87bb20-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.200089 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjtrt\" (UniqueName: \"kubernetes.io/projected/66c9aa39-8cad-4807-98c7-ee329d87bb20-kube-api-access-pjtrt\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.200097 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7ace20-c228-447f-98b7-dc09e562d5b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.200107 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgbf7\" (UniqueName: \"kubernetes.io/projected/aa7ace20-c228-447f-98b7-dc09e562d5b4-kube-api-access-dgbf7\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.200115 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c9aa39-8cad-4807-98c7-ee329d87bb20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.223356 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa7ace20-c228-447f-98b7-dc09e562d5b4-config-data" (OuterVolumeSpecName: "config-data") pod "aa7ace20-c228-447f-98b7-dc09e562d5b4" (UID: "aa7ace20-c228-447f-98b7-dc09e562d5b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.225377 4857 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.306505 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7ace20-c228-447f-98b7-dc09e562d5b4-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.306551 4857 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.500779 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d27f5397-78ae-4289-a153-343f8ed889fa","Type":"ContainerDied","Data":"9d1b17a1e6cb4f6755136ca47182056cb7db92fbacd866a5025db7ca7969777d"} Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.501468 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.536883 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ppkc6" event={"ID":"66c9aa39-8cad-4807-98c7-ee329d87bb20","Type":"ContainerDied","Data":"1a00bbd0dd2057d6365d7bd4d4fb27cd9a597b96554336bde7748886ef43929e"} Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.536960 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a00bbd0dd2057d6365d7bd4d4fb27cd9a597b96554336bde7748886ef43929e" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.537163 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ppkc6" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.556060 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xzd26" event={"ID":"aa7ace20-c228-447f-98b7-dc09e562d5b4","Type":"ContainerDied","Data":"b7095a0b24874e965f6c5c7d07f7cca9c9e975f6f81cef565d89ef17993b5375"} Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.556131 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7095a0b24874e965f6c5c7d07f7cca9c9e975f6f81cef565d89ef17993b5375" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.556290 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xzd26" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.561867 4857 scope.go:117] "RemoveContainer" containerID="eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3" Dec 01 21:55:31 crc kubenswrapper[4857]: E1201 21:55:31.563601 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3\": container with ID starting with eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3 not found: ID does not exist" containerID="eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.563663 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3"} err="failed to get container status \"eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3\": rpc error: code = NotFound desc = could not find container \"eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3\": container with ID starting with eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3 not found: ID does not exist" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.563695 4857 scope.go:117] "RemoveContainer" containerID="448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.563782 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:55:31 crc kubenswrapper[4857]: E1201 21:55:31.566177 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42\": container with ID starting with 448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42 not found: ID does not exist" containerID="448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.566209 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42"} err="failed to get container status \"448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42\": rpc error: code = NotFound desc = could not find container \"448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42\": container with ID starting with 448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42 not found: ID does not exist" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.566227 4857 scope.go:117] "RemoveContainer" containerID="eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.584807 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3"} err="failed to get container status \"eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3\": rpc error: code = NotFound desc = could not find container \"eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3\": container with ID starting with eca10841632c7343ab16280e42672474001b59880bf2d766a3c188075126edd3 not found: ID does not exist" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.584857 4857 scope.go:117] "RemoveContainer" containerID="448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.598566 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.603193 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42"} err="failed to get container status \"448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42\": rpc error: code = NotFound desc = could not find container \"448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42\": container with ID starting with 448db9d05e8c4c813b10aa323f8979ecd17bec9361214e50971a2d8fa8af7f42 not found: ID does not exist" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.603235 4857 scope.go:117] "RemoveContainer" containerID="9df5f41182fb925d21f9486bde5ef4d9bf7bb43ce5b8371d3334430802a937b4" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.621109 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:55:31 crc kubenswrapper[4857]: E1201 21:55:31.621813 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d27f5397-78ae-4289-a153-343f8ed889fa" containerName="glance-log" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.621839 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d27f5397-78ae-4289-a153-343f8ed889fa" containerName="glance-log" Dec 01 21:55:31 crc kubenswrapper[4857]: E1201 21:55:31.621871 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66c9aa39-8cad-4807-98c7-ee329d87bb20" containerName="barbican-db-sync" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.621880 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="66c9aa39-8cad-4807-98c7-ee329d87bb20" containerName="barbican-db-sync" Dec 01 21:55:31 crc kubenswrapper[4857]: E1201 21:55:31.621898 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d27f5397-78ae-4289-a153-343f8ed889fa" containerName="glance-httpd" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.621905 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d27f5397-78ae-4289-a153-343f8ed889fa" containerName="glance-httpd" Dec 01 21:55:31 crc kubenswrapper[4857]: E1201 21:55:31.621929 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7ace20-c228-447f-98b7-dc09e562d5b4" containerName="heat-db-sync" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.621942 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7ace20-c228-447f-98b7-dc09e562d5b4" containerName="heat-db-sync" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.633269 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="d27f5397-78ae-4289-a153-343f8ed889fa" containerName="glance-log" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.633314 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa7ace20-c228-447f-98b7-dc09e562d5b4" containerName="heat-db-sync" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.633332 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="d27f5397-78ae-4289-a153-343f8ed889fa" containerName="glance-httpd" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.633377 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="66c9aa39-8cad-4807-98c7-ee329d87bb20" containerName="barbican-db-sync" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.634949 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.635714 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.638780 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.640525 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.646829 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.730740 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.734170 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07e01fec-a494-44c1-9671-b5f32f79e96a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.734441 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.734649 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07e01fec-a494-44c1-9671-b5f32f79e96a-logs\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.734782 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.734900 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-scripts\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.750086 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-config-data\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.750271 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tswh\" (UniqueName: \"kubernetes.io/projected/07e01fec-a494-44c1-9671-b5f32f79e96a-kube-api-access-6tswh\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.774542 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-k46r8"] Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.774853 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" podUID="6c2f464a-9080-43f7-bf3e-ac92a9aaea77" containerName="dnsmasq-dns" containerID="cri-o://7cb10cb5343476af2d24bdb3e5c30494885537c087187ee67ea2d82bd9a98c52" gracePeriod=10 Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.781470 4857 scope.go:117] "RemoveContainer" containerID="d665290d31a51681d8f8c4e4b0751cb4fb073752e3e30f19fc9b4381f051f96e" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.851479 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.851522 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07e01fec-a494-44c1-9671-b5f32f79e96a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.851574 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.851607 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07e01fec-a494-44c1-9671-b5f32f79e96a-logs\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.851650 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.851683 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-scripts\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.851714 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-config-data\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.851762 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tswh\" (UniqueName: \"kubernetes.io/projected/07e01fec-a494-44c1-9671-b5f32f79e96a-kube-api-access-6tswh\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.853073 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.855867 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07e01fec-a494-44c1-9671-b5f32f79e96a-logs\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.859787 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07e01fec-a494-44c1-9671-b5f32f79e96a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.865226 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d27f5397-78ae-4289-a153-343f8ed889fa" path="/var/lib/kubelet/pods/d27f5397-78ae-4289-a153-343f8ed889fa/volumes" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.874993 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.877728 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.884735 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-scripts\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.892124 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-config-data\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.966080 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:31 crc kubenswrapper[4857]: I1201 21:55:31.982815 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tswh\" (UniqueName: \"kubernetes.io/projected/07e01fec-a494-44c1-9671-b5f32f79e96a-kube-api-access-6tswh\") pod \"glance-default-external-api-0\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " pod="openstack/glance-default-external-api-0" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.281224 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.572667 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-db75d4bb-tjp4b"] Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.574114 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.583768 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.583894 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-894zj" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.584027 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.610618 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5f6bc6688c-77x7p"] Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.612873 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.616565 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.672568 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-db75d4bb-tjp4b"] Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.702124 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d92f4970-9829-4417-9cc1-9436c8993a18-config-data-custom\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.702562 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-logs\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.702581 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-config-data-custom\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.702634 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-combined-ca-bundle\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.702662 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc6jp\" (UniqueName: \"kubernetes.io/projected/d92f4970-9829-4417-9cc1-9436c8993a18-kube-api-access-lc6jp\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.702687 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d92f4970-9829-4417-9cc1-9436c8993a18-combined-ca-bundle\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.702734 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d92f4970-9829-4417-9cc1-9436c8993a18-logs\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.702760 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-config-data\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.702785 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8zn5\" (UniqueName: \"kubernetes.io/projected/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-kube-api-access-h8zn5\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.702802 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d92f4970-9829-4417-9cc1-9436c8993a18-config-data\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.743114 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5f6bc6688c-77x7p"] Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.768628 4857 generic.go:334] "Generic (PLEG): container finished" podID="6c2f464a-9080-43f7-bf3e-ac92a9aaea77" containerID="7cb10cb5343476af2d24bdb3e5c30494885537c087187ee67ea2d82bd9a98c52" exitCode=0 Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.768672 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" event={"ID":"6c2f464a-9080-43f7-bf3e-ac92a9aaea77","Type":"ContainerDied","Data":"7cb10cb5343476af2d24bdb3e5c30494885537c087187ee67ea2d82bd9a98c52"} Dec 01 21:55:32 crc kubenswrapper[4857]: E1201 21:55:32.798300 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.808539 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-combined-ca-bundle\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.808599 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc6jp\" (UniqueName: \"kubernetes.io/projected/d92f4970-9829-4417-9cc1-9436c8993a18-kube-api-access-lc6jp\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.808640 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d92f4970-9829-4417-9cc1-9436c8993a18-combined-ca-bundle\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.808720 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d92f4970-9829-4417-9cc1-9436c8993a18-logs\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.808755 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-config-data\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.808792 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8zn5\" (UniqueName: \"kubernetes.io/projected/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-kube-api-access-h8zn5\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.808813 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d92f4970-9829-4417-9cc1-9436c8993a18-config-data\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.808886 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d92f4970-9829-4417-9cc1-9436c8993a18-config-data-custom\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.808912 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-logs\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.808931 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-config-data-custom\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.821606 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d92f4970-9829-4417-9cc1-9436c8993a18-logs\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: W1201 21:55:32.821727 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ec1f836_90ab_4f44_b500_5a705a165925.slice/crio-1808b55fd2ae872c67b19382b9fba0ef2118c61f9d83f14dfec98339ef26f83d WatchSource:0}: Error finding container 1808b55fd2ae872c67b19382b9fba0ef2118c61f9d83f14dfec98339ef26f83d: Status 404 returned error can't find the container with id 1808b55fd2ae872c67b19382b9fba0ef2118c61f9d83f14dfec98339ef26f83d Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.826807 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-logs\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.838700 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-config-data-custom\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.842650 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-combined-ca-bundle\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.843256 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d92f4970-9829-4417-9cc1-9436c8993a18-config-data\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.845761 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d92f4970-9829-4417-9cc1-9436c8993a18-config-data-custom\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.856418 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d92f4970-9829-4417-9cc1-9436c8993a18-combined-ca-bundle\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.856621 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-mddfv"] Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.858793 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.859602 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-config-data\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.876603 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc6jp\" (UniqueName: \"kubernetes.io/projected/d92f4970-9829-4417-9cc1-9436c8993a18-kube-api-access-lc6jp\") pod \"barbican-keystone-listener-db75d4bb-tjp4b\" (UID: \"d92f4970-9829-4417-9cc1-9436c8993a18\") " pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.911346 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-config\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.911440 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-dns-svc\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.911516 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.911542 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgnf6\" (UniqueName: \"kubernetes.io/projected/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-kube-api-access-qgnf6\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.911573 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.911639 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.914291 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8zn5\" (UniqueName: \"kubernetes.io/projected/5ae783f5-078d-49c0-9b57-cf796b0e8f3b-kube-api-access-h8zn5\") pod \"barbican-worker-5f6bc6688c-77x7p\" (UID: \"5ae783f5-078d-49c0-9b57-cf796b0e8f3b\") " pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.923588 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-mddfv"] Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.925813 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.934543 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-67679ff7fd-jqkpt"] Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.947568 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.949736 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" Dec 01 21:55:32 crc kubenswrapper[4857]: I1201 21:55:32.950429 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.033249 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.033357 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-config-data\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.033415 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-config-data-custom\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.033513 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-config\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.033619 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-combined-ca-bundle\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.033729 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-dns-svc\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.033813 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g6lq\" (UniqueName: \"kubernetes.io/projected/b51fe719-2f9b-49a2-a3b4-3bb083590b02-kube-api-access-9g6lq\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.033858 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b51fe719-2f9b-49a2-a3b4-3bb083590b02-logs\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.034277 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.034310 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgnf6\" (UniqueName: \"kubernetes.io/projected/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-kube-api-access-qgnf6\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.034402 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.044931 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.047931 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5f6bc6688c-77x7p" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.051413 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.053206 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-67679ff7fd-jqkpt"] Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.080277 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-config\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.081491 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-dns-svc\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.089031 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.091027 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgnf6\" (UniqueName: \"kubernetes.io/projected/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-kube-api-access-qgnf6\") pod \"dnsmasq-dns-85ff748b95-mddfv\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.136106 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-combined-ca-bundle\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.136435 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g6lq\" (UniqueName: \"kubernetes.io/projected/b51fe719-2f9b-49a2-a3b4-3bb083590b02-kube-api-access-9g6lq\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.136522 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b51fe719-2f9b-49a2-a3b4-3bb083590b02-logs\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.136656 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-config-data\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.136730 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-config-data-custom\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.142398 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b51fe719-2f9b-49a2-a3b4-3bb083590b02-logs\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.143673 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-combined-ca-bundle\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.148370 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-config-data-custom\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.151172 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-config-data\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.170394 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.178128 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g6lq\" (UniqueName: \"kubernetes.io/projected/b51fe719-2f9b-49a2-a3b4-3bb083590b02-kube-api-access-9g6lq\") pod \"barbican-api-67679ff7fd-jqkpt\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.197280 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.208221 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.241062 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-config\") pod \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.241141 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-dns-swift-storage-0\") pod \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.241278 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-ovsdbserver-sb\") pod \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.241316 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn5xp\" (UniqueName: \"kubernetes.io/projected/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-kube-api-access-sn5xp\") pod \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.241345 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-dns-svc\") pod \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.241419 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-ovsdbserver-nb\") pod \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\" (UID: \"6c2f464a-9080-43f7-bf3e-ac92a9aaea77\") " Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.283552 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-kube-api-access-sn5xp" (OuterVolumeSpecName: "kube-api-access-sn5xp") pod "6c2f464a-9080-43f7-bf3e-ac92a9aaea77" (UID: "6c2f464a-9080-43f7-bf3e-ac92a9aaea77"). InnerVolumeSpecName "kube-api-access-sn5xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.307714 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6c2f464a-9080-43f7-bf3e-ac92a9aaea77" (UID: "6c2f464a-9080-43f7-bf3e-ac92a9aaea77"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.345019 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-config" (OuterVolumeSpecName: "config") pod "6c2f464a-9080-43f7-bf3e-ac92a9aaea77" (UID: "6c2f464a-9080-43f7-bf3e-ac92a9aaea77"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.348016 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.348566 4857 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.348581 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn5xp\" (UniqueName: \"kubernetes.io/projected/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-kube-api-access-sn5xp\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.365670 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6c2f464a-9080-43f7-bf3e-ac92a9aaea77" (UID: "6c2f464a-9080-43f7-bf3e-ac92a9aaea77"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.387273 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6c2f464a-9080-43f7-bf3e-ac92a9aaea77" (UID: "6c2f464a-9080-43f7-bf3e-ac92a9aaea77"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.391006 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6c2f464a-9080-43f7-bf3e-ac92a9aaea77" (UID: "6c2f464a-9080-43f7-bf3e-ac92a9aaea77"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.453334 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.453395 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.453407 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c2f464a-9080-43f7-bf3e-ac92a9aaea77-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.497188 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.720961 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-db75d4bb-tjp4b"] Dec 01 21:55:33 crc kubenswrapper[4857]: W1201 21:55:33.735736 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd92f4970_9829_4417_9cc1_9436c8993a18.slice/crio-9ee9c69ba1e75e6410b140f95f1dfef130aaf7338d812a9f71389df0b8f2905e WatchSource:0}: Error finding container 9ee9c69ba1e75e6410b140f95f1dfef130aaf7338d812a9f71389df0b8f2905e: Status 404 returned error can't find the container with id 9ee9c69ba1e75e6410b140f95f1dfef130aaf7338d812a9f71389df0b8f2905e Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.813176 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" event={"ID":"d92f4970-9829-4417-9cc1-9436c8993a18","Type":"ContainerStarted","Data":"9ee9c69ba1e75e6410b140f95f1dfef130aaf7338d812a9f71389df0b8f2905e"} Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.910498 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3ec1f836-90ab-4f44-b500-5a705a165925","Type":"ContainerStarted","Data":"1808b55fd2ae872c67b19382b9fba0ef2118c61f9d83f14dfec98339ef26f83d"} Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.910536 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-x296d" event={"ID":"dc87da26-cfe1-401c-b115-bab361f3c126","Type":"ContainerStarted","Data":"6e264677551a6c313a6fa79f7941ae8b1ce5ea6b290d42d7ef24c21e8963a6d5"} Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.910552 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.910567 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5f6bc6688c-77x7p"] Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.910581 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57b8cf5fc7-kg4kw" event={"ID":"33d2facf-f0c8-4566-a15c-038f0f9f9a6e","Type":"ContainerStarted","Data":"9c9d801ec453701ca9cf060c58ac47ac1484ee6923c5a81c22c00595f43a8d67"} Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.910590 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57b8cf5fc7-kg4kw" event={"ID":"33d2facf-f0c8-4566-a15c-038f0f9f9a6e","Type":"ContainerStarted","Data":"ade60bc905a84dea21c0260297d0c29ac8073945a03f6539b75820f9c10e3aa8"} Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.915672 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f166659f-d3ef-4f73-8258-46f19eacd6d2","Type":"ContainerStarted","Data":"1316489558d6e4f55dd048e49f02f63980aa08f4393838910026d208c03438eb"} Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.916003 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerName="ceilometer-notification-agent" containerID="cri-o://d41567b7fae5b528ba83d21d75265a44c2e44823455abdf64fc298e6d51f7a33" gracePeriod=30 Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.916611 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerName="sg-core" containerID="cri-o://7936e0e57d464af88b39cfa9e719b2e64386ed51d93bf91ea00cba4b6f4e14b9" gracePeriod=30 Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.916705 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerName="proxy-httpd" containerID="cri-o://1316489558d6e4f55dd048e49f02f63980aa08f4393838910026d208c03438eb" gracePeriod=30 Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.917300 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.929873 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-57b8cf5fc7-kg4kw" podStartSLOduration=9.929849727 podStartE2EDuration="9.929849727s" podCreationTimestamp="2025-12-01 21:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:33.922479117 +0000 UTC m=+1252.412541434" watchObservedRunningTime="2025-12-01 21:55:33.929849727 +0000 UTC m=+1252.419912044" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.937073 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" event={"ID":"6c2f464a-9080-43f7-bf3e-ac92a9aaea77","Type":"ContainerDied","Data":"c042212b472a2246dda7177757dab2916e74011fff85fbacbf92da761523f23a"} Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.937124 4857 scope.go:117] "RemoveContainer" containerID="7cb10cb5343476af2d24bdb3e5c30494885537c087187ee67ea2d82bd9a98c52" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.937236 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-k46r8" Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.972900 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"07e01fec-a494-44c1-9671-b5f32f79e96a","Type":"ContainerStarted","Data":"5cd726cebba3a46175fd824f89fad26e2ac79ed69a4a7746e2b1dd1c08f4628b"} Dec 01 21:55:33 crc kubenswrapper[4857]: I1201 21:55:33.982069 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-x296d" podStartSLOduration=4.196946217 podStartE2EDuration="1m3.982033427s" podCreationTimestamp="2025-12-01 21:54:30 +0000 UTC" firstStartedPulling="2025-12-01 21:54:32.015966188 +0000 UTC m=+1190.506028505" lastFinishedPulling="2025-12-01 21:55:31.801053398 +0000 UTC m=+1250.291115715" observedRunningTime="2025-12-01 21:55:33.970091464 +0000 UTC m=+1252.460153791" watchObservedRunningTime="2025-12-01 21:55:33.982033427 +0000 UTC m=+1252.472095744" Dec 01 21:55:34 crc kubenswrapper[4857]: I1201 21:55:34.037412 4857 scope.go:117] "RemoveContainer" containerID="444cb8be53d41af1d94de563163e451ef5fe4e936480f3c85a517a5ee37df0c4" Dec 01 21:55:34 crc kubenswrapper[4857]: I1201 21:55:34.054911 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-k46r8"] Dec 01 21:55:34 crc kubenswrapper[4857]: I1201 21:55:34.073132 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-k46r8"] Dec 01 21:55:34 crc kubenswrapper[4857]: I1201 21:55:34.113969 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-67679ff7fd-jqkpt"] Dec 01 21:55:34 crc kubenswrapper[4857]: I1201 21:55:34.123161 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-mddfv"] Dec 01 21:55:34 crc kubenswrapper[4857]: W1201 21:55:34.211879 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10703969_5fd7_4a79_a8e6_48ebfbaf98ba.slice/crio-62f99ed0bd0cbd57ba8ba0340fb215169d43d33e2a76ba3d8676de502a051e90 WatchSource:0}: Error finding container 62f99ed0bd0cbd57ba8ba0340fb215169d43d33e2a76ba3d8676de502a051e90: Status 404 returned error can't find the container with id 62f99ed0bd0cbd57ba8ba0340fb215169d43d33e2a76ba3d8676de502a051e90 Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.127765 4857 generic.go:334] "Generic (PLEG): container finished" podID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerID="1316489558d6e4f55dd048e49f02f63980aa08f4393838910026d208c03438eb" exitCode=0 Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.128182 4857 generic.go:334] "Generic (PLEG): container finished" podID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerID="7936e0e57d464af88b39cfa9e719b2e64386ed51d93bf91ea00cba4b6f4e14b9" exitCode=2 Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.128292 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f166659f-d3ef-4f73-8258-46f19eacd6d2","Type":"ContainerDied","Data":"1316489558d6e4f55dd048e49f02f63980aa08f4393838910026d208c03438eb"} Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.128332 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f166659f-d3ef-4f73-8258-46f19eacd6d2","Type":"ContainerDied","Data":"7936e0e57d464af88b39cfa9e719b2e64386ed51d93bf91ea00cba4b6f4e14b9"} Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.206248 4857 generic.go:334] "Generic (PLEG): container finished" podID="10703969-5fd7-4a79-a8e6-48ebfbaf98ba" containerID="16a1c2937e6903f9ac98d09f9b50a6ed18ae7429c00058d0e2e28a724014a3a2" exitCode=0 Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.206592 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" event={"ID":"10703969-5fd7-4a79-a8e6-48ebfbaf98ba","Type":"ContainerDied","Data":"16a1c2937e6903f9ac98d09f9b50a6ed18ae7429c00058d0e2e28a724014a3a2"} Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.217621 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" event={"ID":"10703969-5fd7-4a79-a8e6-48ebfbaf98ba","Type":"ContainerStarted","Data":"62f99ed0bd0cbd57ba8ba0340fb215169d43d33e2a76ba3d8676de502a051e90"} Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.258395 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5f6bc6688c-77x7p" event={"ID":"5ae783f5-078d-49c0-9b57-cf796b0e8f3b","Type":"ContainerStarted","Data":"38d452a6ca0e003fe73c8f83a6b8a321a5dacd035df65ea64e7ae950e22a9a4e"} Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.304008 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"07e01fec-a494-44c1-9671-b5f32f79e96a","Type":"ContainerStarted","Data":"4925b165c7077377e62bea2da2c5c31c88ce758e3a6407de1e148ddf0131b679"} Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.319506 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3ec1f836-90ab-4f44-b500-5a705a165925","Type":"ContainerStarted","Data":"152968eff1600d4b0bb4c27e45c9f98e2a9426bf72c50fed47e7d7d5af868bf2"} Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.329995 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67679ff7fd-jqkpt" event={"ID":"b51fe719-2f9b-49a2-a3b4-3bb083590b02","Type":"ContainerStarted","Data":"426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966"} Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.330111 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67679ff7fd-jqkpt" event={"ID":"b51fe719-2f9b-49a2-a3b4-3bb083590b02","Type":"ContainerStarted","Data":"94b1c3ffc5b0e03f079146451731aa6df4ccd4e72582c82d150fd35ef6bec4a7"} Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.851927 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c2f464a-9080-43f7-bf3e-ac92a9aaea77" path="/var/lib/kubelet/pods/6c2f464a-9080-43f7-bf3e-ac92a9aaea77/volumes" Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.919135 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-cf44fcf6b-c86cx"] Dec 01 21:55:35 crc kubenswrapper[4857]: E1201 21:55:35.919894 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c2f464a-9080-43f7-bf3e-ac92a9aaea77" containerName="dnsmasq-dns" Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.919917 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c2f464a-9080-43f7-bf3e-ac92a9aaea77" containerName="dnsmasq-dns" Dec 01 21:55:35 crc kubenswrapper[4857]: E1201 21:55:35.919931 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c2f464a-9080-43f7-bf3e-ac92a9aaea77" containerName="init" Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.919939 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c2f464a-9080-43f7-bf3e-ac92a9aaea77" containerName="init" Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.920211 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c2f464a-9080-43f7-bf3e-ac92a9aaea77" containerName="dnsmasq-dns" Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.921625 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.923924 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.924376 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 01 21:55:35 crc kubenswrapper[4857]: I1201 21:55:35.934564 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-cf44fcf6b-c86cx"] Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.082525 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-public-tls-certs\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.082616 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-combined-ca-bundle\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.082670 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-internal-tls-certs\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.082784 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-config-data-custom\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.082826 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13134eb9-c4d1-4060-8fd3-266afd46f420-logs\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.082857 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sjgz\" (UniqueName: \"kubernetes.io/projected/13134eb9-c4d1-4060-8fd3-266afd46f420-kube-api-access-8sjgz\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.083050 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-config-data\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.185556 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-public-tls-certs\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.185622 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-combined-ca-bundle\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.185659 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-internal-tls-certs\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.185711 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-config-data-custom\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.185736 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13134eb9-c4d1-4060-8fd3-266afd46f420-logs\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.185767 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sjgz\" (UniqueName: \"kubernetes.io/projected/13134eb9-c4d1-4060-8fd3-266afd46f420-kube-api-access-8sjgz\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.185789 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-config-data\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.187724 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13134eb9-c4d1-4060-8fd3-266afd46f420-logs\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.194810 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-combined-ca-bundle\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.195231 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-internal-tls-certs\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.196695 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-config-data\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.198615 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-config-data-custom\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.227071 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/13134eb9-c4d1-4060-8fd3-266afd46f420-public-tls-certs\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.230848 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sjgz\" (UniqueName: \"kubernetes.io/projected/13134eb9-c4d1-4060-8fd3-266afd46f420-kube-api-access-8sjgz\") pod \"barbican-api-cf44fcf6b-c86cx\" (UID: \"13134eb9-c4d1-4060-8fd3-266afd46f420\") " pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.279675 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.351493 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3ec1f836-90ab-4f44-b500-5a705a165925","Type":"ContainerStarted","Data":"0761be5c409a673752c97c6527eca4e0dabc3ce5b2e9374fa9db1ff742f67ccd"} Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.369567 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67679ff7fd-jqkpt" event={"ID":"b51fe719-2f9b-49a2-a3b4-3bb083590b02","Type":"ContainerStarted","Data":"e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e"} Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.369704 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.389101 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.389082527 podStartE2EDuration="11.389082527s" podCreationTimestamp="2025-12-01 21:55:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:36.381407529 +0000 UTC m=+1254.871469846" watchObservedRunningTime="2025-12-01 21:55:36.389082527 +0000 UTC m=+1254.879144844" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.392611 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" event={"ID":"10703969-5fd7-4a79-a8e6-48ebfbaf98ba","Type":"ContainerStarted","Data":"7ca6fae616ceed3994a91ad5e5a707b43763c18f780375f9cc1aa8f667e1b4fa"} Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.393597 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.400540 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"07e01fec-a494-44c1-9671-b5f32f79e96a","Type":"ContainerStarted","Data":"52086b2db0db0092223027f370cb55c8d6cbdd2a7c93d9296553b1a04101e5db"} Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.498806 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-67679ff7fd-jqkpt" podStartSLOduration=4.498785296 podStartE2EDuration="4.498785296s" podCreationTimestamp="2025-12-01 21:55:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:36.405290984 +0000 UTC m=+1254.895353291" watchObservedRunningTime="2025-12-01 21:55:36.498785296 +0000 UTC m=+1254.988847613" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.524992 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" podStartSLOduration=4.524972778 podStartE2EDuration="4.524972778s" podCreationTimestamp="2025-12-01 21:55:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:36.515874105 +0000 UTC m=+1255.005936422" watchObservedRunningTime="2025-12-01 21:55:36.524972778 +0000 UTC m=+1255.015035095" Dec 01 21:55:36 crc kubenswrapper[4857]: I1201 21:55:36.546347 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.546330321 podStartE2EDuration="5.546330321s" podCreationTimestamp="2025-12-01 21:55:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:36.545286026 +0000 UTC m=+1255.035348343" watchObservedRunningTime="2025-12-01 21:55:36.546330321 +0000 UTC m=+1255.036392638" Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.423966 4857 generic.go:334] "Generic (PLEG): container finished" podID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerID="d41567b7fae5b528ba83d21d75265a44c2e44823455abdf64fc298e6d51f7a33" exitCode=0 Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.424036 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f166659f-d3ef-4f73-8258-46f19eacd6d2","Type":"ContainerDied","Data":"d41567b7fae5b528ba83d21d75265a44c2e44823455abdf64fc298e6d51f7a33"} Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.436127 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.436326 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f166659f-d3ef-4f73-8258-46f19eacd6d2","Type":"ContainerDied","Data":"3647ba61d437efb6da3e37f975093fc94522f182b8150ac50e7f2c060fa3c31e"} Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.436647 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3647ba61d437efb6da3e37f975093fc94522f182b8150ac50e7f2c060fa3c31e" Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.670740 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.837540 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f166659f-d3ef-4f73-8258-46f19eacd6d2-log-httpd\") pod \"f166659f-d3ef-4f73-8258-46f19eacd6d2\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.837690 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2pkf\" (UniqueName: \"kubernetes.io/projected/f166659f-d3ef-4f73-8258-46f19eacd6d2-kube-api-access-b2pkf\") pod \"f166659f-d3ef-4f73-8258-46f19eacd6d2\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.837731 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f166659f-d3ef-4f73-8258-46f19eacd6d2-run-httpd\") pod \"f166659f-d3ef-4f73-8258-46f19eacd6d2\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.837784 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-sg-core-conf-yaml\") pod \"f166659f-d3ef-4f73-8258-46f19eacd6d2\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.837856 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-config-data\") pod \"f166659f-d3ef-4f73-8258-46f19eacd6d2\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.837923 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-scripts\") pod \"f166659f-d3ef-4f73-8258-46f19eacd6d2\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.837965 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-combined-ca-bundle\") pod \"f166659f-d3ef-4f73-8258-46f19eacd6d2\" (UID: \"f166659f-d3ef-4f73-8258-46f19eacd6d2\") " Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.842514 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f166659f-d3ef-4f73-8258-46f19eacd6d2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f166659f-d3ef-4f73-8258-46f19eacd6d2" (UID: "f166659f-d3ef-4f73-8258-46f19eacd6d2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.846528 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f166659f-d3ef-4f73-8258-46f19eacd6d2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f166659f-d3ef-4f73-8258-46f19eacd6d2" (UID: "f166659f-d3ef-4f73-8258-46f19eacd6d2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.846903 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f166659f-d3ef-4f73-8258-46f19eacd6d2-kube-api-access-b2pkf" (OuterVolumeSpecName: "kube-api-access-b2pkf") pod "f166659f-d3ef-4f73-8258-46f19eacd6d2" (UID: "f166659f-d3ef-4f73-8258-46f19eacd6d2"). InnerVolumeSpecName "kube-api-access-b2pkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.857447 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-scripts" (OuterVolumeSpecName: "scripts") pod "f166659f-d3ef-4f73-8258-46f19eacd6d2" (UID: "f166659f-d3ef-4f73-8258-46f19eacd6d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.946273 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.946322 4857 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f166659f-d3ef-4f73-8258-46f19eacd6d2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.946342 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2pkf\" (UniqueName: \"kubernetes.io/projected/f166659f-d3ef-4f73-8258-46f19eacd6d2-kube-api-access-b2pkf\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.946355 4857 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f166659f-d3ef-4f73-8258-46f19eacd6d2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:37 crc kubenswrapper[4857]: I1201 21:55:37.986358 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-cf44fcf6b-c86cx"] Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.100914 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f166659f-d3ef-4f73-8258-46f19eacd6d2" (UID: "f166659f-d3ef-4f73-8258-46f19eacd6d2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.118227 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-config-data" (OuterVolumeSpecName: "config-data") pod "f166659f-d3ef-4f73-8258-46f19eacd6d2" (UID: "f166659f-d3ef-4f73-8258-46f19eacd6d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.152781 4857 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.152821 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.157467 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f166659f-d3ef-4f73-8258-46f19eacd6d2" (UID: "f166659f-d3ef-4f73-8258-46f19eacd6d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.254941 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f166659f-d3ef-4f73-8258-46f19eacd6d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.435215 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cf44fcf6b-c86cx" event={"ID":"13134eb9-c4d1-4060-8fd3-266afd46f420","Type":"ContainerStarted","Data":"f4e70ef1f800fd89e7b4f2bc4c700cc16bd55e5b490cc2fe4ee9570a567a675a"} Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.435263 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cf44fcf6b-c86cx" event={"ID":"13134eb9-c4d1-4060-8fd3-266afd46f420","Type":"ContainerStarted","Data":"94a4d2d746f43207a61154d6c69c7f88cd1826752f8758e32e4a303694eb4dc8"} Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.440228 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5f6bc6688c-77x7p" event={"ID":"5ae783f5-078d-49c0-9b57-cf796b0e8f3b","Type":"ContainerStarted","Data":"b4a213a24448e991d42119c8d2107624110939d0d22c4c34a9cac23032d0d191"} Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.440291 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5f6bc6688c-77x7p" event={"ID":"5ae783f5-078d-49c0-9b57-cf796b0e8f3b","Type":"ContainerStarted","Data":"000ef221117199f024bee55998d0fb6d42ceddfccaafcd494076c705c4127588"} Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.442758 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" event={"ID":"d92f4970-9829-4417-9cc1-9436c8993a18","Type":"ContainerStarted","Data":"865fd50157acd2745f6eb393111abc83f419bda573df41a2495c6389e5989774"} Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.442807 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" event={"ID":"d92f4970-9829-4417-9cc1-9436c8993a18","Type":"ContainerStarted","Data":"4d717ee6700b34d4d64aedf437a0368274896b4d2017df3d7e509d58cd6457b2"} Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.443036 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.484643 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5f6bc6688c-77x7p" podStartSLOduration=2.970440414 podStartE2EDuration="6.484620242s" podCreationTimestamp="2025-12-01 21:55:32 +0000 UTC" firstStartedPulling="2025-12-01 21:55:33.916102091 +0000 UTC m=+1252.406164408" lastFinishedPulling="2025-12-01 21:55:37.430281899 +0000 UTC m=+1255.920344236" observedRunningTime="2025-12-01 21:55:38.469460191 +0000 UTC m=+1256.959522508" watchObservedRunningTime="2025-12-01 21:55:38.484620242 +0000 UTC m=+1256.974682569" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.535287 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-db75d4bb-tjp4b" podStartSLOduration=2.849797037 podStartE2EDuration="6.535263914s" podCreationTimestamp="2025-12-01 21:55:32 +0000 UTC" firstStartedPulling="2025-12-01 21:55:33.738951418 +0000 UTC m=+1252.229013735" lastFinishedPulling="2025-12-01 21:55:37.424418285 +0000 UTC m=+1255.914480612" observedRunningTime="2025-12-01 21:55:38.502358037 +0000 UTC m=+1256.992420374" watchObservedRunningTime="2025-12-01 21:55:38.535263914 +0000 UTC m=+1257.025326231" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.562226 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.580570 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.601139 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:55:38 crc kubenswrapper[4857]: E1201 21:55:38.601831 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerName="sg-core" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.601850 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerName="sg-core" Dec 01 21:55:38 crc kubenswrapper[4857]: E1201 21:55:38.601889 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerName="ceilometer-notification-agent" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.601896 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerName="ceilometer-notification-agent" Dec 01 21:55:38 crc kubenswrapper[4857]: E1201 21:55:38.601920 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerName="proxy-httpd" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.601927 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerName="proxy-httpd" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.602152 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerName="proxy-httpd" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.602170 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerName="ceilometer-notification-agent" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.602185 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" containerName="sg-core" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.604355 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.609644 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.609798 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.615944 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.663411 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp9jr\" (UniqueName: \"kubernetes.io/projected/42bfab31-c501-4a3d-9625-d95c2cb43ffa-kube-api-access-qp9jr\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.663493 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-scripts\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.663569 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42bfab31-c501-4a3d-9625-d95c2cb43ffa-log-httpd\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.663667 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.663835 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42bfab31-c501-4a3d-9625-d95c2cb43ffa-run-httpd\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.663861 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.663927 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-config-data\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.765580 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-config-data\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.765630 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp9jr\" (UniqueName: \"kubernetes.io/projected/42bfab31-c501-4a3d-9625-d95c2cb43ffa-kube-api-access-qp9jr\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.765678 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-scripts\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.765746 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42bfab31-c501-4a3d-9625-d95c2cb43ffa-log-httpd\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.765770 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.765814 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42bfab31-c501-4a3d-9625-d95c2cb43ffa-run-httpd\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.765832 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.767181 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42bfab31-c501-4a3d-9625-d95c2cb43ffa-run-httpd\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.767320 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42bfab31-c501-4a3d-9625-d95c2cb43ffa-log-httpd\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.773943 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-config-data\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.774562 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-scripts\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.781745 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.789102 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.801629 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp9jr\" (UniqueName: \"kubernetes.io/projected/42bfab31-c501-4a3d-9625-d95c2cb43ffa-kube-api-access-qp9jr\") pod \"ceilometer-0\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " pod="openstack/ceilometer-0" Dec 01 21:55:38 crc kubenswrapper[4857]: I1201 21:55:38.937741 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:55:39 crc kubenswrapper[4857]: I1201 21:55:39.455168 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cf44fcf6b-c86cx" event={"ID":"13134eb9-c4d1-4060-8fd3-266afd46f420","Type":"ContainerStarted","Data":"45faf51b746c98f1624dbed81e1680beee43aff46bd7d9d7e0e460d47476db94"} Dec 01 21:55:39 crc kubenswrapper[4857]: I1201 21:55:39.455427 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:39 crc kubenswrapper[4857]: I1201 21:55:39.455448 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:39 crc kubenswrapper[4857]: I1201 21:55:39.497998 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-cf44fcf6b-c86cx" podStartSLOduration=4.497969411 podStartE2EDuration="4.497969411s" podCreationTimestamp="2025-12-01 21:55:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:39.488436208 +0000 UTC m=+1257.978498525" watchObservedRunningTime="2025-12-01 21:55:39.497969411 +0000 UTC m=+1257.988031718" Dec 01 21:55:39 crc kubenswrapper[4857]: I1201 21:55:39.595285 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:55:39 crc kubenswrapper[4857]: I1201 21:55:39.849824 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f166659f-d3ef-4f73-8258-46f19eacd6d2" path="/var/lib/kubelet/pods/f166659f-d3ef-4f73-8258-46f19eacd6d2/volumes" Dec 01 21:55:40 crc kubenswrapper[4857]: I1201 21:55:40.464851 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42bfab31-c501-4a3d-9625-d95c2cb43ffa","Type":"ContainerStarted","Data":"e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2"} Dec 01 21:55:40 crc kubenswrapper[4857]: I1201 21:55:40.465331 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42bfab31-c501-4a3d-9625-d95c2cb43ffa","Type":"ContainerStarted","Data":"d3e2e01c677a5acc95073859a6e571c0513325988d462a329a78d150ed84e96b"} Dec 01 21:55:40 crc kubenswrapper[4857]: I1201 21:55:40.468649 4857 generic.go:334] "Generic (PLEG): container finished" podID="dc87da26-cfe1-401c-b115-bab361f3c126" containerID="6e264677551a6c313a6fa79f7941ae8b1ce5ea6b290d42d7ef24c21e8963a6d5" exitCode=0 Dec 01 21:55:40 crc kubenswrapper[4857]: I1201 21:55:40.468715 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-x296d" event={"ID":"dc87da26-cfe1-401c-b115-bab361f3c126","Type":"ContainerDied","Data":"6e264677551a6c313a6fa79f7941ae8b1ce5ea6b290d42d7ef24c21e8963a6d5"} Dec 01 21:55:41 crc kubenswrapper[4857]: I1201 21:55:41.538373 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42bfab31-c501-4a3d-9625-d95c2cb43ffa","Type":"ContainerStarted","Data":"711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f"} Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.143818 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-x296d" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.246862 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-combined-ca-bundle\") pod \"dc87da26-cfe1-401c-b115-bab361f3c126\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.247107 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dc87da26-cfe1-401c-b115-bab361f3c126-etc-machine-id\") pod \"dc87da26-cfe1-401c-b115-bab361f3c126\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.247146 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-config-data\") pod \"dc87da26-cfe1-401c-b115-bab361f3c126\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.247168 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-scripts\") pod \"dc87da26-cfe1-401c-b115-bab361f3c126\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.247188 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc87da26-cfe1-401c-b115-bab361f3c126-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "dc87da26-cfe1-401c-b115-bab361f3c126" (UID: "dc87da26-cfe1-401c-b115-bab361f3c126"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.247231 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-db-sync-config-data\") pod \"dc87da26-cfe1-401c-b115-bab361f3c126\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.247250 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6pch\" (UniqueName: \"kubernetes.io/projected/dc87da26-cfe1-401c-b115-bab361f3c126-kube-api-access-r6pch\") pod \"dc87da26-cfe1-401c-b115-bab361f3c126\" (UID: \"dc87da26-cfe1-401c-b115-bab361f3c126\") " Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.248899 4857 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dc87da26-cfe1-401c-b115-bab361f3c126-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.256174 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-scripts" (OuterVolumeSpecName: "scripts") pod "dc87da26-cfe1-401c-b115-bab361f3c126" (UID: "dc87da26-cfe1-401c-b115-bab361f3c126"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.258397 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc87da26-cfe1-401c-b115-bab361f3c126-kube-api-access-r6pch" (OuterVolumeSpecName: "kube-api-access-r6pch") pod "dc87da26-cfe1-401c-b115-bab361f3c126" (UID: "dc87da26-cfe1-401c-b115-bab361f3c126"). InnerVolumeSpecName "kube-api-access-r6pch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.270547 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "dc87da26-cfe1-401c-b115-bab361f3c126" (UID: "dc87da26-cfe1-401c-b115-bab361f3c126"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.283904 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.284806 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.289264 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc87da26-cfe1-401c-b115-bab361f3c126" (UID: "dc87da26-cfe1-401c-b115-bab361f3c126"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.325368 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-config-data" (OuterVolumeSpecName: "config-data") pod "dc87da26-cfe1-401c-b115-bab361f3c126" (UID: "dc87da26-cfe1-401c-b115-bab361f3c126"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.325588 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.335231 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.356713 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.356765 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.356779 4857 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.356793 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6pch\" (UniqueName: \"kubernetes.io/projected/dc87da26-cfe1-401c-b115-bab361f3c126-kube-api-access-r6pch\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.356807 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc87da26-cfe1-401c-b115-bab361f3c126-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.565299 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42bfab31-c501-4a3d-9625-d95c2cb43ffa","Type":"ContainerStarted","Data":"f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4"} Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.575962 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-x296d" event={"ID":"dc87da26-cfe1-401c-b115-bab361f3c126","Type":"ContainerDied","Data":"68e1a6ce912af12a209d3643e35a8fe97485121f4f423cf151ac4c9cbc829c05"} Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.576031 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68e1a6ce912af12a209d3643e35a8fe97485121f4f423cf151ac4c9cbc829c05" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.576141 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-x296d" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.577168 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.577203 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.834085 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 21:55:42 crc kubenswrapper[4857]: E1201 21:55:42.834466 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc87da26-cfe1-401c-b115-bab361f3c126" containerName="cinder-db-sync" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.834482 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc87da26-cfe1-401c-b115-bab361f3c126" containerName="cinder-db-sync" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.834666 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc87da26-cfe1-401c-b115-bab361f3c126" containerName="cinder-db-sync" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.835748 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.838668 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.838913 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.839227 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.840646 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-d5kl6" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.888510 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.888750 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-config-data\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.888832 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.888982 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qwc5\" (UniqueName: \"kubernetes.io/projected/2aa3910b-cc16-4514-a74b-8e5061e2ef40-kube-api-access-5qwc5\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.889095 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2aa3910b-cc16-4514-a74b-8e5061e2ef40-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.889574 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-scripts\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.916034 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 21:55:42 crc kubenswrapper[4857]: I1201 21:55:42.995779 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2aa3910b-cc16-4514-a74b-8e5061e2ef40-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:42.993969 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2aa3910b-cc16-4514-a74b-8e5061e2ef40-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.011141 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-scripts\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.011268 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.011412 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-config-data\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.011470 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.011584 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qwc5\" (UniqueName: \"kubernetes.io/projected/2aa3910b-cc16-4514-a74b-8e5061e2ef40-kube-api-access-5qwc5\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.018367 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.023487 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-mddfv"] Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.024013 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" podUID="10703969-5fd7-4a79-a8e6-48ebfbaf98ba" containerName="dnsmasq-dns" containerID="cri-o://7ca6fae616ceed3994a91ad5e5a707b43763c18f780375f9cc1aa8f667e1b4fa" gracePeriod=10 Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.025417 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.028366 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-scripts\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.029679 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-config-data\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.034194 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.045061 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qwc5\" (UniqueName: \"kubernetes.io/projected/2aa3910b-cc16-4514-a74b-8e5061e2ef40-kube-api-access-5qwc5\") pod \"cinder-scheduler-0\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.079342 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-r8rpn"] Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.081287 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.119199 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-r8rpn"] Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.171501 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" podUID="10703969-5fd7-4a79-a8e6-48ebfbaf98ba" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.157:5353: connect: connection refused" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.177762 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.215276 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.215320 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-config\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.215339 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.215444 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.215478 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c59jz\" (UniqueName: \"kubernetes.io/projected/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-kube-api-access-c59jz\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.215504 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.222193 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.227868 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.230843 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.242016 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.318700 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.319145 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da1a9dd7-58b2-458e-8134-8251beec7d02-logs\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.319306 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpd85\" (UniqueName: \"kubernetes.io/projected/da1a9dd7-58b2-458e-8134-8251beec7d02-kube-api-access-vpd85\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.319401 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da1a9dd7-58b2-458e-8134-8251beec7d02-etc-machine-id\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.319485 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-config-data-custom\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.319577 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.319665 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-config\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.319820 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.319913 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-scripts\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.320052 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.320155 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.320311 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-config-data\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.320442 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c59jz\" (UniqueName: \"kubernetes.io/projected/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-kube-api-access-c59jz\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.322064 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.322789 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.323625 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-config\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.324289 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.324908 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.344747 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c59jz\" (UniqueName: \"kubernetes.io/projected/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-kube-api-access-c59jz\") pod \"dnsmasq-dns-5c9776ccc5-r8rpn\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.423601 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.424396 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-config-data\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.424589 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da1a9dd7-58b2-458e-8134-8251beec7d02-logs\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.424710 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpd85\" (UniqueName: \"kubernetes.io/projected/da1a9dd7-58b2-458e-8134-8251beec7d02-kube-api-access-vpd85\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.424781 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da1a9dd7-58b2-458e-8134-8251beec7d02-etc-machine-id\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.424863 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-config-data-custom\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.428781 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-scripts\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.429602 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da1a9dd7-58b2-458e-8134-8251beec7d02-logs\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.429844 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da1a9dd7-58b2-458e-8134-8251beec7d02-etc-machine-id\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.441778 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.441808 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-config-data-custom\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.442433 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-scripts\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.450834 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-config-data\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.453901 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpd85\" (UniqueName: \"kubernetes.io/projected/da1a9dd7-58b2-458e-8134-8251beec7d02-kube-api-access-vpd85\") pod \"cinder-api-0\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.592624 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.630392 4857 generic.go:334] "Generic (PLEG): container finished" podID="10703969-5fd7-4a79-a8e6-48ebfbaf98ba" containerID="7ca6fae616ceed3994a91ad5e5a707b43763c18f780375f9cc1aa8f667e1b4fa" exitCode=0 Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.631360 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" event={"ID":"10703969-5fd7-4a79-a8e6-48ebfbaf98ba","Type":"ContainerDied","Data":"7ca6fae616ceed3994a91ad5e5a707b43763c18f780375f9cc1aa8f667e1b4fa"} Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.710077 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.808651 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.944981 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-config\") pod \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.945482 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-dns-swift-storage-0\") pod \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.945510 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-ovsdbserver-nb\") pod \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.945554 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-ovsdbserver-sb\") pod \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.945617 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-dns-svc\") pod \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.945714 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgnf6\" (UniqueName: \"kubernetes.io/projected/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-kube-api-access-qgnf6\") pod \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\" (UID: \"10703969-5fd7-4a79-a8e6-48ebfbaf98ba\") " Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.953212 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-kube-api-access-qgnf6" (OuterVolumeSpecName: "kube-api-access-qgnf6") pod "10703969-5fd7-4a79-a8e6-48ebfbaf98ba" (UID: "10703969-5fd7-4a79-a8e6-48ebfbaf98ba"). InnerVolumeSpecName "kube-api-access-qgnf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:43 crc kubenswrapper[4857]: I1201 21:55:43.983757 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.048365 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgnf6\" (UniqueName: \"kubernetes.io/projected/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-kube-api-access-qgnf6\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.056188 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "10703969-5fd7-4a79-a8e6-48ebfbaf98ba" (UID: "10703969-5fd7-4a79-a8e6-48ebfbaf98ba"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.082258 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "10703969-5fd7-4a79-a8e6-48ebfbaf98ba" (UID: "10703969-5fd7-4a79-a8e6-48ebfbaf98ba"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.084653 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "10703969-5fd7-4a79-a8e6-48ebfbaf98ba" (UID: "10703969-5fd7-4a79-a8e6-48ebfbaf98ba"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.098515 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "10703969-5fd7-4a79-a8e6-48ebfbaf98ba" (UID: "10703969-5fd7-4a79-a8e6-48ebfbaf98ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.115483 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-config" (OuterVolumeSpecName: "config") pod "10703969-5fd7-4a79-a8e6-48ebfbaf98ba" (UID: "10703969-5fd7-4a79-a8e6-48ebfbaf98ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.153245 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.153710 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.153726 4857 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.153738 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.153762 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10703969-5fd7-4a79-a8e6-48ebfbaf98ba-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.728347 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" event={"ID":"10703969-5fd7-4a79-a8e6-48ebfbaf98ba","Type":"ContainerDied","Data":"62f99ed0bd0cbd57ba8ba0340fb215169d43d33e2a76ba3d8676de502a051e90"} Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.728402 4857 scope.go:117] "RemoveContainer" containerID="7ca6fae616ceed3994a91ad5e5a707b43763c18f780375f9cc1aa8f667e1b4fa" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.728534 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-mddfv" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.757191 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.757221 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.757244 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2aa3910b-cc16-4514-a74b-8e5061e2ef40","Type":"ContainerStarted","Data":"cae89b42148470f8472f6f5b698c8ef30932e9131f26c9ce1d9aaf02278203c8"} Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.814918 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-mddfv"] Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.825257 4857 scope.go:117] "RemoveContainer" containerID="16a1c2937e6903f9ac98d09f9b50a6ed18ae7429c00058d0e2e28a724014a3a2" Dec 01 21:55:44 crc kubenswrapper[4857]: I1201 21:55:44.848152 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-mddfv"] Dec 01 21:55:45 crc kubenswrapper[4857]: I1201 21:55:45.015175 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-r8rpn"] Dec 01 21:55:45 crc kubenswrapper[4857]: I1201 21:55:45.027286 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 21:55:45 crc kubenswrapper[4857]: I1201 21:55:45.824508 4857 generic.go:334] "Generic (PLEG): container finished" podID="bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" containerID="9c57cf46fd8ad499bf7a40af7b25958254da9960c76192fc72491ff863ece35d" exitCode=0 Dec 01 21:55:45 crc kubenswrapper[4857]: I1201 21:55:45.824813 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" event={"ID":"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12","Type":"ContainerDied","Data":"9c57cf46fd8ad499bf7a40af7b25958254da9960c76192fc72491ff863ece35d"} Dec 01 21:55:45 crc kubenswrapper[4857]: I1201 21:55:45.824839 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" event={"ID":"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12","Type":"ContainerStarted","Data":"f37419770688b54e6d2c6c2f0df62d40fbfee22c53d9fbd59923f4e913d3dae8"} Dec 01 21:55:45 crc kubenswrapper[4857]: I1201 21:55:45.874206 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10703969-5fd7-4a79-a8e6-48ebfbaf98ba" path="/var/lib/kubelet/pods/10703969-5fd7-4a79-a8e6-48ebfbaf98ba/volumes" Dec 01 21:55:45 crc kubenswrapper[4857]: I1201 21:55:45.874797 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da1a9dd7-58b2-458e-8134-8251beec7d02","Type":"ContainerStarted","Data":"182b4e022df02796ba488a9af226cbe150441e4f06afd68c36b9b7210036a3fa"} Dec 01 21:55:45 crc kubenswrapper[4857]: I1201 21:55:45.878335 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42bfab31-c501-4a3d-9625-d95c2cb43ffa","Type":"ContainerStarted","Data":"6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db"} Dec 01 21:55:45 crc kubenswrapper[4857]: I1201 21:55:45.879513 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.154230 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.154278 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.200998 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.223596 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.009290216 podStartE2EDuration="8.223580066s" podCreationTimestamp="2025-12-01 21:55:38 +0000 UTC" firstStartedPulling="2025-12-01 21:55:39.66028677 +0000 UTC m=+1258.150349087" lastFinishedPulling="2025-12-01 21:55:44.87457662 +0000 UTC m=+1263.364638937" observedRunningTime="2025-12-01 21:55:45.936896469 +0000 UTC m=+1264.426958786" watchObservedRunningTime="2025-12-01 21:55:46.223580066 +0000 UTC m=+1264.713642383" Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.234552 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.420305 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.872979 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.873501 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.946066 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" event={"ID":"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12","Type":"ContainerStarted","Data":"e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463"} Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.946345 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.952343 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da1a9dd7-58b2-458e-8134-8251beec7d02","Type":"ContainerStarted","Data":"acb2dabf8bafc309791b5cce9c795324445611eb89691058d21620bed518c8c0"} Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.954112 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2aa3910b-cc16-4514-a74b-8e5061e2ef40","Type":"ContainerStarted","Data":"683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820"} Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.956770 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.956794 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:46 crc kubenswrapper[4857]: I1201 21:55:46.983198 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" podStartSLOduration=4.983175855 podStartE2EDuration="4.983175855s" podCreationTimestamp="2025-12-01 21:55:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:46.969629123 +0000 UTC m=+1265.459691440" watchObservedRunningTime="2025-12-01 21:55:46.983175855 +0000 UTC m=+1265.473238162" Dec 01 21:55:47 crc kubenswrapper[4857]: I1201 21:55:47.085311 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 21:55:47 crc kubenswrapper[4857]: I1201 21:55:47.237518 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:47 crc kubenswrapper[4857]: I1201 21:55:47.983265 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da1a9dd7-58b2-458e-8134-8251beec7d02","Type":"ContainerStarted","Data":"b0e084fffedc0dcf7b68c91f7c8bcbbed5b83dcbcbb7b5f088756c5a4cbbd871"} Dec 01 21:55:47 crc kubenswrapper[4857]: I1201 21:55:47.984238 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="da1a9dd7-58b2-458e-8134-8251beec7d02" containerName="cinder-api-log" containerID="cri-o://acb2dabf8bafc309791b5cce9c795324445611eb89691058d21620bed518c8c0" gracePeriod=30 Dec 01 21:55:47 crc kubenswrapper[4857]: I1201 21:55:47.984326 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 21:55:47 crc kubenswrapper[4857]: I1201 21:55:47.984681 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="da1a9dd7-58b2-458e-8134-8251beec7d02" containerName="cinder-api" containerID="cri-o://b0e084fffedc0dcf7b68c91f7c8bcbbed5b83dcbcbb7b5f088756c5a4cbbd871" gracePeriod=30 Dec 01 21:55:48 crc kubenswrapper[4857]: I1201 21:55:48.013716 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.013698714 podStartE2EDuration="5.013698714s" podCreationTimestamp="2025-12-01 21:55:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:48.011998472 +0000 UTC m=+1266.502060789" watchObservedRunningTime="2025-12-01 21:55:48.013698714 +0000 UTC m=+1266.503761031" Dec 01 21:55:48 crc kubenswrapper[4857]: I1201 21:55:48.243321 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-67679ff7fd-jqkpt" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 21:55:49 crc kubenswrapper[4857]: I1201 21:55:49.017131 4857 generic.go:334] "Generic (PLEG): container finished" podID="da1a9dd7-58b2-458e-8134-8251beec7d02" containerID="acb2dabf8bafc309791b5cce9c795324445611eb89691058d21620bed518c8c0" exitCode=143 Dec 01 21:55:49 crc kubenswrapper[4857]: I1201 21:55:49.017224 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da1a9dd7-58b2-458e-8134-8251beec7d02","Type":"ContainerDied","Data":"acb2dabf8bafc309791b5cce9c795324445611eb89691058d21620bed518c8c0"} Dec 01 21:55:49 crc kubenswrapper[4857]: I1201 21:55:49.019944 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2aa3910b-cc16-4514-a74b-8e5061e2ef40","Type":"ContainerStarted","Data":"d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1"} Dec 01 21:55:49 crc kubenswrapper[4857]: I1201 21:55:49.019970 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:55:49 crc kubenswrapper[4857]: I1201 21:55:49.019991 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:55:49 crc kubenswrapper[4857]: I1201 21:55:49.047933 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.728919083 podStartE2EDuration="7.047912874s" podCreationTimestamp="2025-12-01 21:55:42 +0000 UTC" firstStartedPulling="2025-12-01 21:55:43.998229509 +0000 UTC m=+1262.488291826" lastFinishedPulling="2025-12-01 21:55:45.3172233 +0000 UTC m=+1263.807285617" observedRunningTime="2025-12-01 21:55:49.041816815 +0000 UTC m=+1267.531879132" watchObservedRunningTime="2025-12-01 21:55:49.047912874 +0000 UTC m=+1267.537975191" Dec 01 21:55:49 crc kubenswrapper[4857]: I1201 21:55:49.517674 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:49 crc kubenswrapper[4857]: I1201 21:55:49.795062 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:49 crc kubenswrapper[4857]: I1201 21:55:49.998267 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-cf44fcf6b-c86cx" Dec 01 21:55:50 crc kubenswrapper[4857]: I1201 21:55:50.083450 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-67679ff7fd-jqkpt"] Dec 01 21:55:50 crc kubenswrapper[4857]: I1201 21:55:50.083683 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-67679ff7fd-jqkpt" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api-log" containerID="cri-o://426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966" gracePeriod=30 Dec 01 21:55:50 crc kubenswrapper[4857]: I1201 21:55:50.083808 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-67679ff7fd-jqkpt" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api" containerID="cri-o://e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e" gracePeriod=30 Dec 01 21:55:50 crc kubenswrapper[4857]: I1201 21:55:50.100408 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-67679ff7fd-jqkpt" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Dec 01 21:55:50 crc kubenswrapper[4857]: I1201 21:55:50.100795 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-67679ff7fd-jqkpt" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Dec 01 21:55:50 crc kubenswrapper[4857]: I1201 21:55:50.115188 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-67679ff7fd-jqkpt" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Dec 01 21:55:50 crc kubenswrapper[4857]: I1201 21:55:50.445019 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6965455568-h4s4z" Dec 01 21:55:51 crc kubenswrapper[4857]: I1201 21:55:51.038757 4857 generic.go:334] "Generic (PLEG): container finished" podID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerID="426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966" exitCode=143 Dec 01 21:55:51 crc kubenswrapper[4857]: I1201 21:55:51.038971 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67679ff7fd-jqkpt" event={"ID":"b51fe719-2f9b-49a2-a3b4-3bb083590b02","Type":"ContainerDied","Data":"426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966"} Dec 01 21:55:51 crc kubenswrapper[4857]: I1201 21:55:51.056155 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-fdd4494c7-87llf" Dec 01 21:55:51 crc kubenswrapper[4857]: I1201 21:55:51.663461 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:51 crc kubenswrapper[4857]: I1201 21:55:51.663579 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:55:51 crc kubenswrapper[4857]: I1201 21:55:51.666450 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 21:55:51 crc kubenswrapper[4857]: I1201 21:55:51.783691 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.179219 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.379585 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 21:55:53 crc kubenswrapper[4857]: E1201 21:55:53.379972 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10703969-5fd7-4a79-a8e6-48ebfbaf98ba" containerName="dnsmasq-dns" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.379988 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="10703969-5fd7-4a79-a8e6-48ebfbaf98ba" containerName="dnsmasq-dns" Dec 01 21:55:53 crc kubenswrapper[4857]: E1201 21:55:53.380003 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10703969-5fd7-4a79-a8e6-48ebfbaf98ba" containerName="init" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.380010 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="10703969-5fd7-4a79-a8e6-48ebfbaf98ba" containerName="init" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.380200 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="10703969-5fd7-4a79-a8e6-48ebfbaf98ba" containerName="dnsmasq-dns" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.380814 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.390220 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.390952 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-sh2vw" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.391465 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.399395 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.496092 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96057e27-7ce8-4606-8519-2a7cb76f47ab-openstack-config-secret\") pod \"openstackclient\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.496259 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27qd2\" (UniqueName: \"kubernetes.io/projected/96057e27-7ce8-4606-8519-2a7cb76f47ab-kube-api-access-27qd2\") pod \"openstackclient\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.496338 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96057e27-7ce8-4606-8519-2a7cb76f47ab-combined-ca-bundle\") pod \"openstackclient\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.496361 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96057e27-7ce8-4606-8519-2a7cb76f47ab-openstack-config\") pod \"openstackclient\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.508396 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.597241 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.599131 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96057e27-7ce8-4606-8519-2a7cb76f47ab-openstack-config\") pod \"openstackclient\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.599330 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96057e27-7ce8-4606-8519-2a7cb76f47ab-openstack-config-secret\") pod \"openstackclient\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.599390 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27qd2\" (UniqueName: \"kubernetes.io/projected/96057e27-7ce8-4606-8519-2a7cb76f47ab-kube-api-access-27qd2\") pod \"openstackclient\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.599435 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96057e27-7ce8-4606-8519-2a7cb76f47ab-combined-ca-bundle\") pod \"openstackclient\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.600101 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96057e27-7ce8-4606-8519-2a7cb76f47ab-openstack-config\") pod \"openstackclient\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.613647 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96057e27-7ce8-4606-8519-2a7cb76f47ab-combined-ca-bundle\") pod \"openstackclient\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.628775 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96057e27-7ce8-4606-8519-2a7cb76f47ab-openstack-config-secret\") pod \"openstackclient\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.634290 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27qd2\" (UniqueName: \"kubernetes.io/projected/96057e27-7ce8-4606-8519-2a7cb76f47ab-kube-api-access-27qd2\") pod \"openstackclient\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.663582 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-mvhhh"] Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.663878 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" podUID="6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" containerName="dnsmasq-dns" containerID="cri-o://4ee1494f162df42894e78241882217b883d847b8735ae5641691eb6d19e7548a" gracePeriod=10 Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.728521 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.728988 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.744472 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.800401 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.828761 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.867282 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.942139 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6821156d-2d04-4608-a5c7-c70b1058fdc0-openstack-config-secret\") pod \"openstackclient\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.942190 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6821156d-2d04-4608-a5c7-c70b1058fdc0-openstack-config\") pod \"openstackclient\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.942260 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6821156d-2d04-4608-a5c7-c70b1058fdc0-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " pod="openstack/openstackclient" Dec 01 21:55:53 crc kubenswrapper[4857]: I1201 21:55:53.942290 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfq9q\" (UniqueName: \"kubernetes.io/projected/6821156d-2d04-4608-a5c7-c70b1058fdc0-kube-api-access-bfq9q\") pod \"openstackclient\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " pod="openstack/openstackclient" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.047417 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6821156d-2d04-4608-a5c7-c70b1058fdc0-openstack-config-secret\") pod \"openstackclient\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " pod="openstack/openstackclient" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.047495 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6821156d-2d04-4608-a5c7-c70b1058fdc0-openstack-config\") pod \"openstackclient\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " pod="openstack/openstackclient" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.047569 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6821156d-2d04-4608-a5c7-c70b1058fdc0-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " pod="openstack/openstackclient" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.047597 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfq9q\" (UniqueName: \"kubernetes.io/projected/6821156d-2d04-4608-a5c7-c70b1058fdc0-kube-api-access-bfq9q\") pod \"openstackclient\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " pod="openstack/openstackclient" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.048708 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6821156d-2d04-4608-a5c7-c70b1058fdc0-openstack-config\") pod \"openstackclient\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " pod="openstack/openstackclient" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.054435 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6821156d-2d04-4608-a5c7-c70b1058fdc0-openstack-config-secret\") pod \"openstackclient\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " pod="openstack/openstackclient" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.071664 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6821156d-2d04-4608-a5c7-c70b1058fdc0-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " pod="openstack/openstackclient" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.074879 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfq9q\" (UniqueName: \"kubernetes.io/projected/6821156d-2d04-4608-a5c7-c70b1058fdc0-kube-api-access-bfq9q\") pod \"openstackclient\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " pod="openstack/openstackclient" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.102378 4857 generic.go:334] "Generic (PLEG): container finished" podID="6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" containerID="4ee1494f162df42894e78241882217b883d847b8735ae5641691eb6d19e7548a" exitCode=0 Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.104137 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" event={"ID":"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a","Type":"ContainerDied","Data":"4ee1494f162df42894e78241882217b883d847b8735ae5641691eb6d19e7548a"} Dec 01 21:55:54 crc kubenswrapper[4857]: E1201 21:55:54.202337 4857 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 01 21:55:54 crc kubenswrapper[4857]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_96057e27-7ce8-4606-8519-2a7cb76f47ab_0(ebf35fa0b3646f3022b7c1d1e2c9481c40856dc64e123c0685a2b6c5a7d11c0f): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ebf35fa0b3646f3022b7c1d1e2c9481c40856dc64e123c0685a2b6c5a7d11c0f" Netns:"/var/run/netns/e940a138-0797-4181-af97-cbf44b699105" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=ebf35fa0b3646f3022b7c1d1e2c9481c40856dc64e123c0685a2b6c5a7d11c0f;K8S_POD_UID=96057e27-7ce8-4606-8519-2a7cb76f47ab" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/96057e27-7ce8-4606-8519-2a7cb76f47ab]: expected pod UID "96057e27-7ce8-4606-8519-2a7cb76f47ab" but got "6821156d-2d04-4608-a5c7-c70b1058fdc0" from Kube API Dec 01 21:55:54 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 21:55:54 crc kubenswrapper[4857]: > Dec 01 21:55:54 crc kubenswrapper[4857]: E1201 21:55:54.202423 4857 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 01 21:55:54 crc kubenswrapper[4857]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_96057e27-7ce8-4606-8519-2a7cb76f47ab_0(ebf35fa0b3646f3022b7c1d1e2c9481c40856dc64e123c0685a2b6c5a7d11c0f): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ebf35fa0b3646f3022b7c1d1e2c9481c40856dc64e123c0685a2b6c5a7d11c0f" Netns:"/var/run/netns/e940a138-0797-4181-af97-cbf44b699105" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=ebf35fa0b3646f3022b7c1d1e2c9481c40856dc64e123c0685a2b6c5a7d11c0f;K8S_POD_UID=96057e27-7ce8-4606-8519-2a7cb76f47ab" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/96057e27-7ce8-4606-8519-2a7cb76f47ab]: expected pod UID "96057e27-7ce8-4606-8519-2a7cb76f47ab" but got "6821156d-2d04-4608-a5c7-c70b1058fdc0" from Kube API Dec 01 21:55:54 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 21:55:54 crc kubenswrapper[4857]: > pod="openstack/openstackclient" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.213122 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.354389 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.404198 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.577453 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-config\") pod \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.577685 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-ovsdbserver-sb\") pod \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.577727 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-dns-svc\") pod \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.577755 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-dns-swift-storage-0\") pod \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.577843 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-ovsdbserver-nb\") pod \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.577918 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2dfn\" (UniqueName: \"kubernetes.io/projected/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-kube-api-access-f2dfn\") pod \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\" (UID: \"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a\") " Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.592445 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-kube-api-access-f2dfn" (OuterVolumeSpecName: "kube-api-access-f2dfn") pod "6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" (UID: "6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a"). InnerVolumeSpecName "kube-api-access-f2dfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.612267 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-67679ff7fd-jqkpt" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:49924->10.217.0.158:9311: read: connection reset by peer" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.626509 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-57b8cf5fc7-kg4kw" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.629802 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-67679ff7fd-jqkpt" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:49914->10.217.0.158:9311: read: connection reset by peer" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.635599 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-67679ff7fd-jqkpt" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": dial tcp 10.217.0.158:9311: connect: connection refused" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.662806 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" (UID: "6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.687362 4857 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.687784 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2dfn\" (UniqueName: \"kubernetes.io/projected/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-kube-api-access-f2dfn\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.735827 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" (UID: "6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.739236 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-config" (OuterVolumeSpecName: "config") pod "6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" (UID: "6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.773592 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" (UID: "6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.777680 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7cf98975f6-8bj6t"] Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.778654 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7cf98975f6-8bj6t" podUID="e04bffff-693d-4909-914d-1e2f784ca47d" containerName="neutron-httpd" containerID="cri-o://95e8a45ce13de4d4d028b19ce5c61af5facfe7703e3996a78f3f2f220ccc1303" gracePeriod=30 Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.778853 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7cf98975f6-8bj6t" podUID="e04bffff-693d-4909-914d-1e2f784ca47d" containerName="neutron-api" containerID="cri-o://c22bf24c46ff5fd423c0066118d5059901c95219f85fd49109d8fe88fe75264d" gracePeriod=30 Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.803998 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" (UID: "6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.808133 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.808168 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.808179 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:54 crc kubenswrapper[4857]: I1201 21:55:54.808189 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.121449 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.121598 4857 generic.go:334] "Generic (PLEG): container finished" podID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerID="e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e" exitCode=0 Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.121725 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67679ff7fd-jqkpt" event={"ID":"b51fe719-2f9b-49a2-a3b4-3bb083590b02","Type":"ContainerDied","Data":"e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e"} Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.121769 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67679ff7fd-jqkpt" event={"ID":"b51fe719-2f9b-49a2-a3b4-3bb083590b02","Type":"ContainerDied","Data":"94b1c3ffc5b0e03f079146451731aa6df4ccd4e72582c82d150fd35ef6bec4a7"} Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.121796 4857 scope.go:117] "RemoveContainer" containerID="e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.126661 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" event={"ID":"6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a","Type":"ContainerDied","Data":"d0232fd272fcc87c94d9c72349d92cbb28bdbed8ef25a02502d3612f379e25b5"} Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.126766 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-mvhhh" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.156578 4857 generic.go:334] "Generic (PLEG): container finished" podID="e04bffff-693d-4909-914d-1e2f784ca47d" containerID="95e8a45ce13de4d4d028b19ce5c61af5facfe7703e3996a78f3f2f220ccc1303" exitCode=0 Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.156845 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="2aa3910b-cc16-4514-a74b-8e5061e2ef40" containerName="cinder-scheduler" containerID="cri-o://683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820" gracePeriod=30 Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.156966 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cf98975f6-8bj6t" event={"ID":"e04bffff-693d-4909-914d-1e2f784ca47d","Type":"ContainerDied","Data":"95e8a45ce13de4d4d028b19ce5c61af5facfe7703e3996a78f3f2f220ccc1303"} Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.157026 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.157754 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="2aa3910b-cc16-4514-a74b-8e5061e2ef40" containerName="probe" containerID="cri-o://d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1" gracePeriod=30 Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.182838 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-mvhhh"] Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.188433 4857 scope.go:117] "RemoveContainer" containerID="426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.190773 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-mvhhh"] Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.198668 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.206472 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="96057e27-7ce8-4606-8519-2a7cb76f47ab" podUID="6821156d-2d04-4608-a5c7-c70b1058fdc0" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.244531 4857 scope.go:117] "RemoveContainer" containerID="e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e" Dec 01 21:55:55 crc kubenswrapper[4857]: E1201 21:55:55.245219 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e\": container with ID starting with e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e not found: ID does not exist" containerID="e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.245280 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e"} err="failed to get container status \"e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e\": rpc error: code = NotFound desc = could not find container \"e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e\": container with ID starting with e04e3fda903a96da8547caceb39cf22a27022bd1829c8a972844ca9d8ff6d88e not found: ID does not exist" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.245328 4857 scope.go:117] "RemoveContainer" containerID="426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966" Dec 01 21:55:55 crc kubenswrapper[4857]: E1201 21:55:55.245785 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966\": container with ID starting with 426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966 not found: ID does not exist" containerID="426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.245804 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966"} err="failed to get container status \"426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966\": rpc error: code = NotFound desc = could not find container \"426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966\": container with ID starting with 426bfafbc23b723c354eaf11a76c1ab789504fefd44aefe8ca70aeec147d8966 not found: ID does not exist" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.245820 4857 scope.go:117] "RemoveContainer" containerID="4ee1494f162df42894e78241882217b883d847b8735ae5641691eb6d19e7548a" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.248261 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.270299 4857 scope.go:117] "RemoveContainer" containerID="325d96063eaaed29781d6ebcbce77f4b3fb0988161519e923c1d12aaad24f6e3" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.322441 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b51fe719-2f9b-49a2-a3b4-3bb083590b02-logs\") pod \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.322924 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27qd2\" (UniqueName: \"kubernetes.io/projected/96057e27-7ce8-4606-8519-2a7cb76f47ab-kube-api-access-27qd2\") pod \"96057e27-7ce8-4606-8519-2a7cb76f47ab\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.322996 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g6lq\" (UniqueName: \"kubernetes.io/projected/b51fe719-2f9b-49a2-a3b4-3bb083590b02-kube-api-access-9g6lq\") pod \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.323158 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b51fe719-2f9b-49a2-a3b4-3bb083590b02-logs" (OuterVolumeSpecName: "logs") pod "b51fe719-2f9b-49a2-a3b4-3bb083590b02" (UID: "b51fe719-2f9b-49a2-a3b4-3bb083590b02"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.323177 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96057e27-7ce8-4606-8519-2a7cb76f47ab-openstack-config-secret\") pod \"96057e27-7ce8-4606-8519-2a7cb76f47ab\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.323282 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-config-data-custom\") pod \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.323366 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-config-data\") pod \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.323411 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-combined-ca-bundle\") pod \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\" (UID: \"b51fe719-2f9b-49a2-a3b4-3bb083590b02\") " Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.323444 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96057e27-7ce8-4606-8519-2a7cb76f47ab-openstack-config\") pod \"96057e27-7ce8-4606-8519-2a7cb76f47ab\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.323464 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96057e27-7ce8-4606-8519-2a7cb76f47ab-combined-ca-bundle\") pod \"96057e27-7ce8-4606-8519-2a7cb76f47ab\" (UID: \"96057e27-7ce8-4606-8519-2a7cb76f47ab\") " Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.324609 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b51fe719-2f9b-49a2-a3b4-3bb083590b02-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.324880 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96057e27-7ce8-4606-8519-2a7cb76f47ab-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "96057e27-7ce8-4606-8519-2a7cb76f47ab" (UID: "96057e27-7ce8-4606-8519-2a7cb76f47ab"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.327616 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96057e27-7ce8-4606-8519-2a7cb76f47ab-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "96057e27-7ce8-4606-8519-2a7cb76f47ab" (UID: "96057e27-7ce8-4606-8519-2a7cb76f47ab"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.329321 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96057e27-7ce8-4606-8519-2a7cb76f47ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96057e27-7ce8-4606-8519-2a7cb76f47ab" (UID: "96057e27-7ce8-4606-8519-2a7cb76f47ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.330576 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96057e27-7ce8-4606-8519-2a7cb76f47ab-kube-api-access-27qd2" (OuterVolumeSpecName: "kube-api-access-27qd2") pod "96057e27-7ce8-4606-8519-2a7cb76f47ab" (UID: "96057e27-7ce8-4606-8519-2a7cb76f47ab"). InnerVolumeSpecName "kube-api-access-27qd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.331326 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b51fe719-2f9b-49a2-a3b4-3bb083590b02-kube-api-access-9g6lq" (OuterVolumeSpecName: "kube-api-access-9g6lq") pod "b51fe719-2f9b-49a2-a3b4-3bb083590b02" (UID: "b51fe719-2f9b-49a2-a3b4-3bb083590b02"). InnerVolumeSpecName "kube-api-access-9g6lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.337303 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b51fe719-2f9b-49a2-a3b4-3bb083590b02" (UID: "b51fe719-2f9b-49a2-a3b4-3bb083590b02"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.359212 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b51fe719-2f9b-49a2-a3b4-3bb083590b02" (UID: "b51fe719-2f9b-49a2-a3b4-3bb083590b02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.379715 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-config-data" (OuterVolumeSpecName: "config-data") pod "b51fe719-2f9b-49a2-a3b4-3bb083590b02" (UID: "b51fe719-2f9b-49a2-a3b4-3bb083590b02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.426310 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27qd2\" (UniqueName: \"kubernetes.io/projected/96057e27-7ce8-4606-8519-2a7cb76f47ab-kube-api-access-27qd2\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.426343 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g6lq\" (UniqueName: \"kubernetes.io/projected/b51fe719-2f9b-49a2-a3b4-3bb083590b02-kube-api-access-9g6lq\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.426352 4857 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96057e27-7ce8-4606-8519-2a7cb76f47ab-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.426365 4857 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.426377 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.426387 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51fe719-2f9b-49a2-a3b4-3bb083590b02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.426396 4857 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96057e27-7ce8-4606-8519-2a7cb76f47ab-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.426405 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96057e27-7ce8-4606-8519-2a7cb76f47ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.857136 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" path="/var/lib/kubelet/pods/6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a/volumes" Dec 01 21:55:55 crc kubenswrapper[4857]: I1201 21:55:55.857948 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96057e27-7ce8-4606-8519-2a7cb76f47ab" path="/var/lib/kubelet/pods/96057e27-7ce8-4606-8519-2a7cb76f47ab/volumes" Dec 01 21:55:56 crc kubenswrapper[4857]: I1201 21:55:56.217709 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6821156d-2d04-4608-a5c7-c70b1058fdc0","Type":"ContainerStarted","Data":"ccd012f001941fe72913947d7ff4292c607068d4b5eb875074c8f37d0d7c7625"} Dec 01 21:55:56 crc kubenswrapper[4857]: I1201 21:55:56.231392 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 21:55:56 crc kubenswrapper[4857]: I1201 21:55:56.231444 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67679ff7fd-jqkpt" Dec 01 21:55:56 crc kubenswrapper[4857]: I1201 21:55:56.250403 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="96057e27-7ce8-4606-8519-2a7cb76f47ab" podUID="6821156d-2d04-4608-a5c7-c70b1058fdc0" Dec 01 21:55:56 crc kubenswrapper[4857]: I1201 21:55:56.296341 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-67679ff7fd-jqkpt"] Dec 01 21:55:56 crc kubenswrapper[4857]: I1201 21:55:56.305822 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-67679ff7fd-jqkpt"] Dec 01 21:55:56 crc kubenswrapper[4857]: I1201 21:55:56.687853 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.253247 4857 generic.go:334] "Generic (PLEG): container finished" podID="2aa3910b-cc16-4514-a74b-8e5061e2ef40" containerID="d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1" exitCode=0 Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.253298 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2aa3910b-cc16-4514-a74b-8e5061e2ef40","Type":"ContainerDied","Data":"d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1"} Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.340201 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-56454df68d-6lxzh"] Dec 01 21:55:57 crc kubenswrapper[4857]: E1201 21:55:57.340806 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" containerName="dnsmasq-dns" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.340831 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" containerName="dnsmasq-dns" Dec 01 21:55:57 crc kubenswrapper[4857]: E1201 21:55:57.340854 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" containerName="init" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.340863 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" containerName="init" Dec 01 21:55:57 crc kubenswrapper[4857]: E1201 21:55:57.340881 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api-log" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.340888 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api-log" Dec 01 21:55:57 crc kubenswrapper[4857]: E1201 21:55:57.340902 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.340908 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.341194 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b25fa6b-9fa4-44f3-8316-ae6dd248aa9a" containerName="dnsmasq-dns" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.341212 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.341226 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" containerName="barbican-api-log" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.341971 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.351645 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-djd7n" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.351686 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.351816 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.362781 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-56454df68d-6lxzh"] Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.389462 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-config-data-custom\") pod \"heat-engine-56454df68d-6lxzh\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.389541 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-combined-ca-bundle\") pod \"heat-engine-56454df68d-6lxzh\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.389570 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j67zx\" (UniqueName: \"kubernetes.io/projected/2842d110-9e75-4911-a86e-29327b79b8f8-kube-api-access-j67zx\") pod \"heat-engine-56454df68d-6lxzh\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.389612 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-config-data\") pod \"heat-engine-56454df68d-6lxzh\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.460310 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-75f4dbf68b-t5t5b"] Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.471321 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.480823 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.495158 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grnrb\" (UniqueName: \"kubernetes.io/projected/69afe439-86e6-490f-9476-ebb89b540192-kube-api-access-grnrb\") pod \"heat-cfnapi-75f4dbf68b-t5t5b\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.495250 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-config-data-custom\") pod \"heat-cfnapi-75f4dbf68b-t5t5b\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.495288 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-config-data-custom\") pod \"heat-engine-56454df68d-6lxzh\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.495316 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-config-data\") pod \"heat-cfnapi-75f4dbf68b-t5t5b\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.495380 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-combined-ca-bundle\") pod \"heat-engine-56454df68d-6lxzh\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.495411 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j67zx\" (UniqueName: \"kubernetes.io/projected/2842d110-9e75-4911-a86e-29327b79b8f8-kube-api-access-j67zx\") pod \"heat-engine-56454df68d-6lxzh\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.495449 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-combined-ca-bundle\") pod \"heat-cfnapi-75f4dbf68b-t5t5b\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.495487 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-config-data\") pod \"heat-engine-56454df68d-6lxzh\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.505235 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-75f4dbf68b-t5t5b"] Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.507999 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-combined-ca-bundle\") pod \"heat-engine-56454df68d-6lxzh\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.508100 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-config-data-custom\") pod \"heat-engine-56454df68d-6lxzh\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.508840 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-config-data\") pod \"heat-engine-56454df68d-6lxzh\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.533911 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j67zx\" (UniqueName: \"kubernetes.io/projected/2842d110-9e75-4911-a86e-29327b79b8f8-kube-api-access-j67zx\") pod \"heat-engine-56454df68d-6lxzh\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.541972 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-jr4rg"] Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.543983 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.600009 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-jr4rg"] Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.600877 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-config-data-custom\") pod \"heat-cfnapi-75f4dbf68b-t5t5b\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.600914 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-config-data\") pod \"heat-cfnapi-75f4dbf68b-t5t5b\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.601010 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-combined-ca-bundle\") pod \"heat-cfnapi-75f4dbf68b-t5t5b\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.601033 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.601092 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.601125 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-config\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.601144 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.601236 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdhkn\" (UniqueName: \"kubernetes.io/projected/c76a0a21-8d25-4168-be04-941041bce44a-kube-api-access-mdhkn\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.601283 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grnrb\" (UniqueName: \"kubernetes.io/projected/69afe439-86e6-490f-9476-ebb89b540192-kube-api-access-grnrb\") pod \"heat-cfnapi-75f4dbf68b-t5t5b\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.601337 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.619327 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-config-data-custom\") pod \"heat-cfnapi-75f4dbf68b-t5t5b\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.622383 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-combined-ca-bundle\") pod \"heat-cfnapi-75f4dbf68b-t5t5b\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.622402 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-config-data\") pod \"heat-cfnapi-75f4dbf68b-t5t5b\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.625445 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grnrb\" (UniqueName: \"kubernetes.io/projected/69afe439-86e6-490f-9476-ebb89b540192-kube-api-access-grnrb\") pod \"heat-cfnapi-75f4dbf68b-t5t5b\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.662507 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-77bf9494c9-4bdzk"] Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.665106 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.679118 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-77bf9494c9-4bdzk"] Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.682540 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.689496 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.705250 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-combined-ca-bundle\") pod \"heat-api-77bf9494c9-4bdzk\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.705432 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.705552 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-config-data\") pod \"heat-api-77bf9494c9-4bdzk\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.705608 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.705689 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-config\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.705738 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.705937 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xmzf\" (UniqueName: \"kubernetes.io/projected/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-kube-api-access-9xmzf\") pod \"heat-api-77bf9494c9-4bdzk\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.705998 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-config-data-custom\") pod \"heat-api-77bf9494c9-4bdzk\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.706057 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdhkn\" (UniqueName: \"kubernetes.io/projected/c76a0a21-8d25-4168-be04-941041bce44a-kube-api-access-mdhkn\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.706338 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.706741 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.707093 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.707955 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.708954 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.723614 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-config\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.759538 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdhkn\" (UniqueName: \"kubernetes.io/projected/c76a0a21-8d25-4168-be04-941041bce44a-kube-api-access-mdhkn\") pod \"dnsmasq-dns-7756b9d78c-jr4rg\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.807830 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-config-data\") pod \"heat-api-77bf9494c9-4bdzk\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.807921 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xmzf\" (UniqueName: \"kubernetes.io/projected/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-kube-api-access-9xmzf\") pod \"heat-api-77bf9494c9-4bdzk\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.807944 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-config-data-custom\") pod \"heat-api-77bf9494c9-4bdzk\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.808228 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-combined-ca-bundle\") pod \"heat-api-77bf9494c9-4bdzk\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.809410 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.809461 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.809502 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.810318 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.810507 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae0326e43bb280b472fd2b09c89e53e5301501138efae77ad81d099184e7a849"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.810558 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://ae0326e43bb280b472fd2b09c89e53e5301501138efae77ad81d099184e7a849" gracePeriod=600 Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.816187 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-combined-ca-bundle\") pod \"heat-api-77bf9494c9-4bdzk\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.818090 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-config-data-custom\") pod \"heat-api-77bf9494c9-4bdzk\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.822294 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-config-data\") pod \"heat-api-77bf9494c9-4bdzk\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.852482 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xmzf\" (UniqueName: \"kubernetes.io/projected/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-kube-api-access-9xmzf\") pod \"heat-api-77bf9494c9-4bdzk\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.887188 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b51fe719-2f9b-49a2-a3b4-3bb083590b02" path="/var/lib/kubelet/pods/b51fe719-2f9b-49a2-a3b4-3bb083590b02/volumes" Dec 01 21:55:57 crc kubenswrapper[4857]: I1201 21:55:57.913876 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:55:58 crc kubenswrapper[4857]: I1201 21:55:58.142657 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:55:58 crc kubenswrapper[4857]: I1201 21:55:58.366559 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="ae0326e43bb280b472fd2b09c89e53e5301501138efae77ad81d099184e7a849" exitCode=0 Dec 01 21:55:58 crc kubenswrapper[4857]: I1201 21:55:58.367954 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"ae0326e43bb280b472fd2b09c89e53e5301501138efae77ad81d099184e7a849"} Dec 01 21:55:58 crc kubenswrapper[4857]: I1201 21:55:58.368279 4857 scope.go:117] "RemoveContainer" containerID="1936180283cd1643268f62d5199bdc5d47710e031991bbd7a33aafc43d3ee862" Dec 01 21:55:58 crc kubenswrapper[4857]: I1201 21:55:58.455071 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-56454df68d-6lxzh"] Dec 01 21:55:58 crc kubenswrapper[4857]: I1201 21:55:58.604563 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-jr4rg"] Dec 01 21:55:58 crc kubenswrapper[4857]: W1201 21:55:58.617489 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc76a0a21_8d25_4168_be04_941041bce44a.slice/crio-987f0ec2088516459cb591cc0aa07da579fd57acf4034c96946cfc5f88a75d7c WatchSource:0}: Error finding container 987f0ec2088516459cb591cc0aa07da579fd57acf4034c96946cfc5f88a75d7c: Status 404 returned error can't find the container with id 987f0ec2088516459cb591cc0aa07da579fd57acf4034c96946cfc5f88a75d7c Dec 01 21:55:58 crc kubenswrapper[4857]: I1201 21:55:58.669582 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-75f4dbf68b-t5t5b"] Dec 01 21:55:58 crc kubenswrapper[4857]: W1201 21:55:58.683904 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69afe439_86e6_490f_9476_ebb89b540192.slice/crio-bd8dc4423f6c5c2cd256329cf4bd0ca73544dccbfa97eac950a60d5c366c17bf WatchSource:0}: Error finding container bd8dc4423f6c5c2cd256329cf4bd0ca73544dccbfa97eac950a60d5c366c17bf: Status 404 returned error can't find the container with id bd8dc4423f6c5c2cd256329cf4bd0ca73544dccbfa97eac950a60d5c366c17bf Dec 01 21:55:58 crc kubenswrapper[4857]: I1201 21:55:58.954094 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-77bf9494c9-4bdzk"] Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.264727 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.360930 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-scripts\") pod \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.361125 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2aa3910b-cc16-4514-a74b-8e5061e2ef40-etc-machine-id\") pod \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.361226 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qwc5\" (UniqueName: \"kubernetes.io/projected/2aa3910b-cc16-4514-a74b-8e5061e2ef40-kube-api-access-5qwc5\") pod \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.361342 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-combined-ca-bundle\") pod \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.361430 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-config-data-custom\") pod \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.361481 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-config-data\") pod \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\" (UID: \"2aa3910b-cc16-4514-a74b-8e5061e2ef40\") " Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.361400 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2aa3910b-cc16-4514-a74b-8e5061e2ef40-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2aa3910b-cc16-4514-a74b-8e5061e2ef40" (UID: "2aa3910b-cc16-4514-a74b-8e5061e2ef40"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.362307 4857 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2aa3910b-cc16-4514-a74b-8e5061e2ef40-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.368066 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2aa3910b-cc16-4514-a74b-8e5061e2ef40" (UID: "2aa3910b-cc16-4514-a74b-8e5061e2ef40"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.368689 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aa3910b-cc16-4514-a74b-8e5061e2ef40-kube-api-access-5qwc5" (OuterVolumeSpecName: "kube-api-access-5qwc5") pod "2aa3910b-cc16-4514-a74b-8e5061e2ef40" (UID: "2aa3910b-cc16-4514-a74b-8e5061e2ef40"). InnerVolumeSpecName "kube-api-access-5qwc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.369562 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-scripts" (OuterVolumeSpecName: "scripts") pod "2aa3910b-cc16-4514-a74b-8e5061e2ef40" (UID: "2aa3910b-cc16-4514-a74b-8e5061e2ef40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.380082 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77bf9494c9-4bdzk" event={"ID":"502899ed-eb9a-4eae-a3f3-1ef7369c08e7","Type":"ContainerStarted","Data":"1820300bc5e769e194e42ec372a079e75bd1f5d88d4b389121ae6b15d8295769"} Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.381505 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" event={"ID":"69afe439-86e6-490f-9476-ebb89b540192","Type":"ContainerStarted","Data":"bd8dc4423f6c5c2cd256329cf4bd0ca73544dccbfa97eac950a60d5c366c17bf"} Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.386387 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"022dacb43671109561ed218fe31f14692c5b447676b12b6fcdbccdc338c9650c"} Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.388651 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-56454df68d-6lxzh" event={"ID":"2842d110-9e75-4911-a86e-29327b79b8f8","Type":"ContainerStarted","Data":"621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b"} Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.388860 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.388889 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-56454df68d-6lxzh" event={"ID":"2842d110-9e75-4911-a86e-29327b79b8f8","Type":"ContainerStarted","Data":"60c988026da607ec25e6c00365966a28d06b75048b35d258cee2d8a1e71cdb71"} Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.394251 4857 generic.go:334] "Generic (PLEG): container finished" podID="2aa3910b-cc16-4514-a74b-8e5061e2ef40" containerID="683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820" exitCode=0 Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.394333 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2aa3910b-cc16-4514-a74b-8e5061e2ef40","Type":"ContainerDied","Data":"683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820"} Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.394385 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2aa3910b-cc16-4514-a74b-8e5061e2ef40","Type":"ContainerDied","Data":"cae89b42148470f8472f6f5b698c8ef30932e9131f26c9ce1d9aaf02278203c8"} Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.394335 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.394409 4857 scope.go:117] "RemoveContainer" containerID="d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.400189 4857 generic.go:334] "Generic (PLEG): container finished" podID="c76a0a21-8d25-4168-be04-941041bce44a" containerID="7f814e0af9b78ad7d49afe09cdda869976239459e3508514ec5392487809be88" exitCode=0 Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.400233 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" event={"ID":"c76a0a21-8d25-4168-be04-941041bce44a","Type":"ContainerDied","Data":"7f814e0af9b78ad7d49afe09cdda869976239459e3508514ec5392487809be88"} Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.400255 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" event={"ID":"c76a0a21-8d25-4168-be04-941041bce44a","Type":"ContainerStarted","Data":"987f0ec2088516459cb591cc0aa07da579fd57acf4034c96946cfc5f88a75d7c"} Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.418557 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-56454df68d-6lxzh" podStartSLOduration=2.418538354 podStartE2EDuration="2.418538354s" podCreationTimestamp="2025-12-01 21:55:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:55:59.410593559 +0000 UTC m=+1277.900655896" watchObservedRunningTime="2025-12-01 21:55:59.418538354 +0000 UTC m=+1277.908600671" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.448400 4857 scope.go:117] "RemoveContainer" containerID="683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.463823 4857 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.463845 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.463856 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qwc5\" (UniqueName: \"kubernetes.io/projected/2aa3910b-cc16-4514-a74b-8e5061e2ef40-kube-api-access-5qwc5\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.497414 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2aa3910b-cc16-4514-a74b-8e5061e2ef40" (UID: "2aa3910b-cc16-4514-a74b-8e5061e2ef40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.542420 4857 scope.go:117] "RemoveContainer" containerID="d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1" Dec 01 21:55:59 crc kubenswrapper[4857]: E1201 21:55:59.547408 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1\": container with ID starting with d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1 not found: ID does not exist" containerID="d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.547449 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1"} err="failed to get container status \"d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1\": rpc error: code = NotFound desc = could not find container \"d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1\": container with ID starting with d945371a0fc21b500fb64f576860a4818b5ea5e624f8c2998188e124d7bd66c1 not found: ID does not exist" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.547471 4857 scope.go:117] "RemoveContainer" containerID="683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.547429 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-config-data" (OuterVolumeSpecName: "config-data") pod "2aa3910b-cc16-4514-a74b-8e5061e2ef40" (UID: "2aa3910b-cc16-4514-a74b-8e5061e2ef40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:55:59 crc kubenswrapper[4857]: E1201 21:55:59.552733 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820\": container with ID starting with 683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820 not found: ID does not exist" containerID="683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.552779 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820"} err="failed to get container status \"683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820\": rpc error: code = NotFound desc = could not find container \"683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820\": container with ID starting with 683f6bf34000e18bd5a64569855599bdf4bb8da789fa621962b5135e16cec820 not found: ID does not exist" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.566954 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.567628 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aa3910b-cc16-4514-a74b-8e5061e2ef40-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.775369 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.789755 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.827735 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 21:55:59 crc kubenswrapper[4857]: E1201 21:55:59.828268 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aa3910b-cc16-4514-a74b-8e5061e2ef40" containerName="cinder-scheduler" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.828285 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aa3910b-cc16-4514-a74b-8e5061e2ef40" containerName="cinder-scheduler" Dec 01 21:55:59 crc kubenswrapper[4857]: E1201 21:55:59.828313 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aa3910b-cc16-4514-a74b-8e5061e2ef40" containerName="probe" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.828321 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aa3910b-cc16-4514-a74b-8e5061e2ef40" containerName="probe" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.828568 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aa3910b-cc16-4514-a74b-8e5061e2ef40" containerName="cinder-scheduler" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.828593 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aa3910b-cc16-4514-a74b-8e5061e2ef40" containerName="probe" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.829924 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.837173 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.856145 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aa3910b-cc16-4514-a74b-8e5061e2ef40" path="/var/lib/kubelet/pods/2aa3910b-cc16-4514-a74b-8e5061e2ef40/volumes" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.858435 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.978466 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.978941 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.979002 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl594\" (UniqueName: \"kubernetes.io/projected/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-kube-api-access-wl594\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.979065 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-scripts\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.979092 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-config-data\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:55:59 crc kubenswrapper[4857]: I1201 21:55:59.979305 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.081303 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl594\" (UniqueName: \"kubernetes.io/projected/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-kube-api-access-wl594\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.081364 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-scripts\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.081389 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-config-data\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.081444 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.081516 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.081532 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.082389 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.090151 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.092113 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-scripts\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.092942 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-config-data\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.104349 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl594\" (UniqueName: \"kubernetes.io/projected/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-kube-api-access-wl594\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.110699 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f38eac29-d14a-4bf8-bbd8-df39c591ee6a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f38eac29-d14a-4bf8-bbd8-df39c591ee6a\") " pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.168236 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.451790 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" event={"ID":"c76a0a21-8d25-4168-be04-941041bce44a","Type":"ContainerStarted","Data":"557d60b969ae3ff9ad723c2be3a36ab7a1c599c3765e9e6b0b8aca470ed7cfee"} Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.452303 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.794591 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" podStartSLOduration=3.7945717329999997 podStartE2EDuration="3.794571733s" podCreationTimestamp="2025-12-01 21:55:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:00.491637927 +0000 UTC m=+1278.981700244" watchObservedRunningTime="2025-12-01 21:56:00.794571733 +0000 UTC m=+1279.284634050" Dec 01 21:56:00 crc kubenswrapper[4857]: I1201 21:56:00.811282 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 21:56:00 crc kubenswrapper[4857]: W1201 21:56:00.831658 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf38eac29_d14a_4bf8_bbd8_df39c591ee6a.slice/crio-24b20106aae79da430624c0558d5715aee4c1479589e9ab5b458bc890b12ca3f WatchSource:0}: Error finding container 24b20106aae79da430624c0558d5715aee4c1479589e9ab5b458bc890b12ca3f: Status 404 returned error can't find the container with id 24b20106aae79da430624c0558d5715aee4c1479589e9ab5b458bc890b12ca3f Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.528827 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f38eac29-d14a-4bf8-bbd8-df39c591ee6a","Type":"ContainerStarted","Data":"24b20106aae79da430624c0558d5715aee4c1479589e9ab5b458bc890b12ca3f"} Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.580277 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7b8476f857-zrr4q"] Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.583197 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b8476f857-zrr4q"] Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.583496 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.587894 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.588778 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.588880 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.620749 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b43470bb-ee08-4170-94c6-4a3acbbce50e-combined-ca-bundle\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.620824 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43470bb-ee08-4170-94c6-4a3acbbce50e-internal-tls-certs\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.620883 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b4hd\" (UniqueName: \"kubernetes.io/projected/b43470bb-ee08-4170-94c6-4a3acbbce50e-kube-api-access-8b4hd\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.620918 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b43470bb-ee08-4170-94c6-4a3acbbce50e-log-httpd\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.620954 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43470bb-ee08-4170-94c6-4a3acbbce50e-public-tls-certs\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.620990 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b43470bb-ee08-4170-94c6-4a3acbbce50e-config-data\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.621077 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b43470bb-ee08-4170-94c6-4a3acbbce50e-run-httpd\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.621118 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b43470bb-ee08-4170-94c6-4a3acbbce50e-etc-swift\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.723362 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b4hd\" (UniqueName: \"kubernetes.io/projected/b43470bb-ee08-4170-94c6-4a3acbbce50e-kube-api-access-8b4hd\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.723437 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b43470bb-ee08-4170-94c6-4a3acbbce50e-log-httpd\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.723462 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43470bb-ee08-4170-94c6-4a3acbbce50e-public-tls-certs\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.723560 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b43470bb-ee08-4170-94c6-4a3acbbce50e-config-data\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.723613 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b43470bb-ee08-4170-94c6-4a3acbbce50e-run-httpd\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.723636 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b43470bb-ee08-4170-94c6-4a3acbbce50e-etc-swift\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.723676 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b43470bb-ee08-4170-94c6-4a3acbbce50e-combined-ca-bundle\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.723706 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43470bb-ee08-4170-94c6-4a3acbbce50e-internal-tls-certs\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.723887 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b43470bb-ee08-4170-94c6-4a3acbbce50e-log-httpd\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.724173 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b43470bb-ee08-4170-94c6-4a3acbbce50e-run-httpd\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.731868 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b43470bb-ee08-4170-94c6-4a3acbbce50e-config-data\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.733006 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b43470bb-ee08-4170-94c6-4a3acbbce50e-combined-ca-bundle\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.733939 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43470bb-ee08-4170-94c6-4a3acbbce50e-public-tls-certs\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.745754 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b4hd\" (UniqueName: \"kubernetes.io/projected/b43470bb-ee08-4170-94c6-4a3acbbce50e-kube-api-access-8b4hd\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.750288 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b43470bb-ee08-4170-94c6-4a3acbbce50e-etc-swift\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.758694 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43470bb-ee08-4170-94c6-4a3acbbce50e-internal-tls-certs\") pod \"swift-proxy-7b8476f857-zrr4q\" (UID: \"b43470bb-ee08-4170-94c6-4a3acbbce50e\") " pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:01 crc kubenswrapper[4857]: I1201 21:56:01.910211 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:03 crc kubenswrapper[4857]: I1201 21:56:03.270463 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:56:03 crc kubenswrapper[4857]: I1201 21:56:03.270993 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="ceilometer-central-agent" containerID="cri-o://e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2" gracePeriod=30 Dec 01 21:56:03 crc kubenswrapper[4857]: I1201 21:56:03.271312 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="proxy-httpd" containerID="cri-o://6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db" gracePeriod=30 Dec 01 21:56:03 crc kubenswrapper[4857]: I1201 21:56:03.271355 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="sg-core" containerID="cri-o://f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4" gracePeriod=30 Dec 01 21:56:03 crc kubenswrapper[4857]: I1201 21:56:03.271391 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="ceilometer-notification-agent" containerID="cri-o://711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f" gracePeriod=30 Dec 01 21:56:03 crc kubenswrapper[4857]: I1201 21:56:03.287671 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 01 21:56:03 crc kubenswrapper[4857]: I1201 21:56:03.673065 4857 generic.go:334] "Generic (PLEG): container finished" podID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerID="6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db" exitCode=0 Dec 01 21:56:03 crc kubenswrapper[4857]: I1201 21:56:03.673592 4857 generic.go:334] "Generic (PLEG): container finished" podID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerID="f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4" exitCode=2 Dec 01 21:56:03 crc kubenswrapper[4857]: I1201 21:56:03.673615 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42bfab31-c501-4a3d-9625-d95c2cb43ffa","Type":"ContainerDied","Data":"6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db"} Dec 01 21:56:03 crc kubenswrapper[4857]: I1201 21:56:03.673642 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42bfab31-c501-4a3d-9625-d95c2cb43ffa","Type":"ContainerDied","Data":"f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4"} Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.280618 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b8476f857-zrr4q"] Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.683589 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b8476f857-zrr4q" event={"ID":"b43470bb-ee08-4170-94c6-4a3acbbce50e","Type":"ContainerStarted","Data":"3081cd97cb904969698f6a2913d0826201d700c7139cd606347a7fc3eac9e693"} Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.699635 4857 generic.go:334] "Generic (PLEG): container finished" podID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerID="e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2" exitCode=0 Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.699704 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42bfab31-c501-4a3d-9625-d95c2cb43ffa","Type":"ContainerDied","Data":"e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2"} Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.710897 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77bf9494c9-4bdzk" event={"ID":"502899ed-eb9a-4eae-a3f3-1ef7369c08e7","Type":"ContainerStarted","Data":"b97342b2bd102c51865c503ee965467bd9493b0f742e791191d394191444f65b"} Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.710992 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.721656 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" event={"ID":"69afe439-86e6-490f-9476-ebb89b540192","Type":"ContainerStarted","Data":"5739d985f701aa8a38cd890c6e351dcaf126790bc2552d67e58b2f5f879831e2"} Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.722583 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.734744 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-77bf9494c9-4bdzk" podStartSLOduration=3.307928055 podStartE2EDuration="7.734728182s" podCreationTimestamp="2025-12-01 21:55:57 +0000 UTC" firstStartedPulling="2025-12-01 21:55:58.96797383 +0000 UTC m=+1277.458036137" lastFinishedPulling="2025-12-01 21:56:03.394773957 +0000 UTC m=+1281.884836264" observedRunningTime="2025-12-01 21:56:04.733635305 +0000 UTC m=+1283.223697612" watchObservedRunningTime="2025-12-01 21:56:04.734728182 +0000 UTC m=+1283.224790499" Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.776680 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" podStartSLOduration=3.071409877 podStartE2EDuration="7.77666073s" podCreationTimestamp="2025-12-01 21:55:57 +0000 UTC" firstStartedPulling="2025-12-01 21:55:58.687829373 +0000 UTC m=+1277.177891690" lastFinishedPulling="2025-12-01 21:56:03.393080226 +0000 UTC m=+1281.883142543" observedRunningTime="2025-12-01 21:56:04.775239845 +0000 UTC m=+1283.265302162" watchObservedRunningTime="2025-12-01 21:56:04.77666073 +0000 UTC m=+1283.266723047" Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.816271 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7c64685c4b-rgnfp"] Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.824059 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.854480 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7c64685c4b-rgnfp"] Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.878254 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-66cd8df7d6-zlb8c"] Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.879648 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.929509 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9fd2d2f-01df-4aa4-bc83-46f30fb0c592-config-data-custom\") pod \"heat-engine-7c64685c4b-rgnfp\" (UID: \"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592\") " pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.929561 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-595fq\" (UniqueName: \"kubernetes.io/projected/d9fd2d2f-01df-4aa4-bc83-46f30fb0c592-kube-api-access-595fq\") pod \"heat-engine-7c64685c4b-rgnfp\" (UID: \"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592\") " pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.929627 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9fd2d2f-01df-4aa4-bc83-46f30fb0c592-combined-ca-bundle\") pod \"heat-engine-7c64685c4b-rgnfp\" (UID: \"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592\") " pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.929685 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9fd2d2f-01df-4aa4-bc83-46f30fb0c592-config-data\") pod \"heat-engine-7c64685c4b-rgnfp\" (UID: \"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592\") " pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.933514 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-f87845c79-xkcmc"] Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.939543 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.955563 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-66cd8df7d6-zlb8c"] Dec 01 21:56:04 crc kubenswrapper[4857]: I1201 21:56:04.966244 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-f87845c79-xkcmc"] Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.032490 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-config-data-custom\") pod \"heat-api-f87845c79-xkcmc\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.032609 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9fd2d2f-01df-4aa4-bc83-46f30fb0c592-combined-ca-bundle\") pod \"heat-engine-7c64685c4b-rgnfp\" (UID: \"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592\") " pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.032699 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9fd2d2f-01df-4aa4-bc83-46f30fb0c592-config-data\") pod \"heat-engine-7c64685c4b-rgnfp\" (UID: \"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592\") " pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.032736 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-config-data-custom\") pod \"heat-cfnapi-66cd8df7d6-zlb8c\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.032793 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-combined-ca-bundle\") pod \"heat-api-f87845c79-xkcmc\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.032823 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-combined-ca-bundle\") pod \"heat-cfnapi-66cd8df7d6-zlb8c\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.032891 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-config-data\") pod \"heat-api-f87845c79-xkcmc\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.032924 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwwv7\" (UniqueName: \"kubernetes.io/projected/e8524783-2889-48f9-b8f9-0d5fb3db15e5-kube-api-access-mwwv7\") pod \"heat-api-f87845c79-xkcmc\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.032949 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-config-data\") pod \"heat-cfnapi-66cd8df7d6-zlb8c\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.032976 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9fd2d2f-01df-4aa4-bc83-46f30fb0c592-config-data-custom\") pod \"heat-engine-7c64685c4b-rgnfp\" (UID: \"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592\") " pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.033011 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58m6t\" (UniqueName: \"kubernetes.io/projected/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-kube-api-access-58m6t\") pod \"heat-cfnapi-66cd8df7d6-zlb8c\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.033034 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-595fq\" (UniqueName: \"kubernetes.io/projected/d9fd2d2f-01df-4aa4-bc83-46f30fb0c592-kube-api-access-595fq\") pod \"heat-engine-7c64685c4b-rgnfp\" (UID: \"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592\") " pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.040379 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9fd2d2f-01df-4aa4-bc83-46f30fb0c592-config-data-custom\") pod \"heat-engine-7c64685c4b-rgnfp\" (UID: \"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592\") " pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.040911 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9fd2d2f-01df-4aa4-bc83-46f30fb0c592-combined-ca-bundle\") pod \"heat-engine-7c64685c4b-rgnfp\" (UID: \"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592\") " pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.048807 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9fd2d2f-01df-4aa4-bc83-46f30fb0c592-config-data\") pod \"heat-engine-7c64685c4b-rgnfp\" (UID: \"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592\") " pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.065963 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-595fq\" (UniqueName: \"kubernetes.io/projected/d9fd2d2f-01df-4aa4-bc83-46f30fb0c592-kube-api-access-595fq\") pod \"heat-engine-7c64685c4b-rgnfp\" (UID: \"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592\") " pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.136310 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-config-data\") pod \"heat-api-f87845c79-xkcmc\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.136373 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwwv7\" (UniqueName: \"kubernetes.io/projected/e8524783-2889-48f9-b8f9-0d5fb3db15e5-kube-api-access-mwwv7\") pod \"heat-api-f87845c79-xkcmc\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.136404 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-config-data\") pod \"heat-cfnapi-66cd8df7d6-zlb8c\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.136435 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58m6t\" (UniqueName: \"kubernetes.io/projected/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-kube-api-access-58m6t\") pod \"heat-cfnapi-66cd8df7d6-zlb8c\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.136503 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-config-data-custom\") pod \"heat-api-f87845c79-xkcmc\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.136601 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-config-data-custom\") pod \"heat-cfnapi-66cd8df7d6-zlb8c\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.136652 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-combined-ca-bundle\") pod \"heat-api-f87845c79-xkcmc\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.136687 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-combined-ca-bundle\") pod \"heat-cfnapi-66cd8df7d6-zlb8c\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.146704 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-config-data-custom\") pod \"heat-api-f87845c79-xkcmc\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.148115 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-config-data\") pod \"heat-cfnapi-66cd8df7d6-zlb8c\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.149183 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-combined-ca-bundle\") pod \"heat-api-f87845c79-xkcmc\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.152067 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-config-data-custom\") pod \"heat-cfnapi-66cd8df7d6-zlb8c\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.161792 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwwv7\" (UniqueName: \"kubernetes.io/projected/e8524783-2889-48f9-b8f9-0d5fb3db15e5-kube-api-access-mwwv7\") pod \"heat-api-f87845c79-xkcmc\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.161981 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-config-data\") pod \"heat-api-f87845c79-xkcmc\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.164220 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-combined-ca-bundle\") pod \"heat-cfnapi-66cd8df7d6-zlb8c\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.167663 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58m6t\" (UniqueName: \"kubernetes.io/projected/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-kube-api-access-58m6t\") pod \"heat-cfnapi-66cd8df7d6-zlb8c\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.241943 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.257342 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.266454 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.832627 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b8476f857-zrr4q" event={"ID":"b43470bb-ee08-4170-94c6-4a3acbbce50e","Type":"ContainerStarted","Data":"e9fb65ce447284c3c385bea540370248f0d8deb4b2bf79281777cbc4cc66a2be"} Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.833067 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b8476f857-zrr4q" event={"ID":"b43470bb-ee08-4170-94c6-4a3acbbce50e","Type":"ContainerStarted","Data":"efdc5e3feb6a7f7bea2215d24d16af326f0a6080a3eb9518a9a72ef827b25f60"} Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.834131 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.834409 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.837088 4857 generic.go:334] "Generic (PLEG): container finished" podID="e04bffff-693d-4909-914d-1e2f784ca47d" containerID="c22bf24c46ff5fd423c0066118d5059901c95219f85fd49109d8fe88fe75264d" exitCode=0 Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.912183 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cf98975f6-8bj6t" event={"ID":"e04bffff-693d-4909-914d-1e2f784ca47d","Type":"ContainerDied","Data":"c22bf24c46ff5fd423c0066118d5059901c95219f85fd49109d8fe88fe75264d"} Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.912230 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f38eac29-d14a-4bf8-bbd8-df39c591ee6a","Type":"ContainerStarted","Data":"c61f314d6d4bebddb8e8a52cd52c3ad92ccc443a32b8f83ebb607e5a777e5780"} Dec 01 21:56:05 crc kubenswrapper[4857]: I1201 21:56:05.912175 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7b8476f857-zrr4q" podStartSLOduration=4.912153952 podStartE2EDuration="4.912153952s" podCreationTimestamp="2025-12-01 21:56:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:05.880895106 +0000 UTC m=+1284.370957423" watchObservedRunningTime="2025-12-01 21:56:05.912153952 +0000 UTC m=+1284.402216269" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.066391 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.170191 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q69kp\" (UniqueName: \"kubernetes.io/projected/e04bffff-693d-4909-914d-1e2f784ca47d-kube-api-access-q69kp\") pod \"e04bffff-693d-4909-914d-1e2f784ca47d\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.170275 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-config\") pod \"e04bffff-693d-4909-914d-1e2f784ca47d\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.170306 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-combined-ca-bundle\") pod \"e04bffff-693d-4909-914d-1e2f784ca47d\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.170378 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-ovndb-tls-certs\") pod \"e04bffff-693d-4909-914d-1e2f784ca47d\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.170479 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-httpd-config\") pod \"e04bffff-693d-4909-914d-1e2f784ca47d\" (UID: \"e04bffff-693d-4909-914d-1e2f784ca47d\") " Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.176806 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e04bffff-693d-4909-914d-1e2f784ca47d" (UID: "e04bffff-693d-4909-914d-1e2f784ca47d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.181058 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e04bffff-693d-4909-914d-1e2f784ca47d-kube-api-access-q69kp" (OuterVolumeSpecName: "kube-api-access-q69kp") pod "e04bffff-693d-4909-914d-1e2f784ca47d" (UID: "e04bffff-693d-4909-914d-1e2f784ca47d"). InnerVolumeSpecName "kube-api-access-q69kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.276195 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-config" (OuterVolumeSpecName: "config") pod "e04bffff-693d-4909-914d-1e2f784ca47d" (UID: "e04bffff-693d-4909-914d-1e2f784ca47d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.277466 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.277505 4857 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.277515 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q69kp\" (UniqueName: \"kubernetes.io/projected/e04bffff-693d-4909-914d-1e2f784ca47d-kube-api-access-q69kp\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.280903 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e04bffff-693d-4909-914d-1e2f784ca47d" (UID: "e04bffff-693d-4909-914d-1e2f784ca47d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.294063 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e04bffff-693d-4909-914d-1e2f784ca47d" (UID: "e04bffff-693d-4909-914d-1e2f784ca47d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.380402 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.380440 4857 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e04bffff-693d-4909-914d-1e2f784ca47d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.502185 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7c64685c4b-rgnfp"] Dec 01 21:56:06 crc kubenswrapper[4857]: W1201 21:56:06.512909 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9fd2d2f_01df_4aa4_bc83_46f30fb0c592.slice/crio-911e645b8c8c207b31ffbae83c5715ac3e5323ff426c8e90f5686f211397fe34 WatchSource:0}: Error finding container 911e645b8c8c207b31ffbae83c5715ac3e5323ff426c8e90f5686f211397fe34: Status 404 returned error can't find the container with id 911e645b8c8c207b31ffbae83c5715ac3e5323ff426c8e90f5686f211397fe34 Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.515089 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-f87845c79-xkcmc"] Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.575615 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-66cd8df7d6-zlb8c"] Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.886581 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f87845c79-xkcmc" event={"ID":"e8524783-2889-48f9-b8f9-0d5fb3db15e5","Type":"ContainerStarted","Data":"8efb290de50b48a4ca1ab7c6cca746b424c7a75463781c6111a8bb3709280477"} Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.890119 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" event={"ID":"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24","Type":"ContainerStarted","Data":"053443615a3a777f1e3150c36f425d5ee7177b6b4ef7fd4e32213fed40524944"} Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.897731 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cf98975f6-8bj6t" event={"ID":"e04bffff-693d-4909-914d-1e2f784ca47d","Type":"ContainerDied","Data":"87669aaec6ee97132b534dc4cb15df194c0ffa37824141d4493d15dd3792e69e"} Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.897899 4857 scope.go:117] "RemoveContainer" containerID="95e8a45ce13de4d4d028b19ce5c61af5facfe7703e3996a78f3f2f220ccc1303" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.897974 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cf98975f6-8bj6t" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.917461 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f38eac29-d14a-4bf8-bbd8-df39c591ee6a","Type":"ContainerStarted","Data":"4e215948b16a5f778a246dd37470e22a41d977e34eb4f555de62ff0de9b2bd8d"} Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.920455 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7c64685c4b-rgnfp" event={"ID":"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592","Type":"ContainerStarted","Data":"911e645b8c8c207b31ffbae83c5715ac3e5323ff426c8e90f5686f211397fe34"} Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.946647 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.946628999 podStartE2EDuration="7.946628999s" podCreationTimestamp="2025-12-01 21:55:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:06.942876257 +0000 UTC m=+1285.432938574" watchObservedRunningTime="2025-12-01 21:56:06.946628999 +0000 UTC m=+1285.436691316" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.965318 4857 scope.go:117] "RemoveContainer" containerID="c22bf24c46ff5fd423c0066118d5059901c95219f85fd49109d8fe88fe75264d" Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.977593 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7cf98975f6-8bj6t"] Dec 01 21:56:06 crc kubenswrapper[4857]: I1201 21:56:06.984660 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7cf98975f6-8bj6t"] Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.330110 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.381543 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-config-data\") pod \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.381584 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42bfab31-c501-4a3d-9625-d95c2cb43ffa-run-httpd\") pod \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.381634 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-scripts\") pod \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.381665 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-sg-core-conf-yaml\") pod \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.381688 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qp9jr\" (UniqueName: \"kubernetes.io/projected/42bfab31-c501-4a3d-9625-d95c2cb43ffa-kube-api-access-qp9jr\") pod \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.381778 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42bfab31-c501-4a3d-9625-d95c2cb43ffa-log-httpd\") pod \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.381802 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-combined-ca-bundle\") pod \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\" (UID: \"42bfab31-c501-4a3d-9625-d95c2cb43ffa\") " Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.389616 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42bfab31-c501-4a3d-9625-d95c2cb43ffa-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "42bfab31-c501-4a3d-9625-d95c2cb43ffa" (UID: "42bfab31-c501-4a3d-9625-d95c2cb43ffa"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.392341 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42bfab31-c501-4a3d-9625-d95c2cb43ffa-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "42bfab31-c501-4a3d-9625-d95c2cb43ffa" (UID: "42bfab31-c501-4a3d-9625-d95c2cb43ffa"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.392760 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42bfab31-c501-4a3d-9625-d95c2cb43ffa-kube-api-access-qp9jr" (OuterVolumeSpecName: "kube-api-access-qp9jr") pod "42bfab31-c501-4a3d-9625-d95c2cb43ffa" (UID: "42bfab31-c501-4a3d-9625-d95c2cb43ffa"). InnerVolumeSpecName "kube-api-access-qp9jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.397843 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-scripts" (OuterVolumeSpecName: "scripts") pod "42bfab31-c501-4a3d-9625-d95c2cb43ffa" (UID: "42bfab31-c501-4a3d-9625-d95c2cb43ffa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.471275 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "42bfab31-c501-4a3d-9625-d95c2cb43ffa" (UID: "42bfab31-c501-4a3d-9625-d95c2cb43ffa"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.500800 4857 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42bfab31-c501-4a3d-9625-d95c2cb43ffa-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.500847 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.500859 4857 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.500871 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qp9jr\" (UniqueName: \"kubernetes.io/projected/42bfab31-c501-4a3d-9625-d95c2cb43ffa-kube-api-access-qp9jr\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.500884 4857 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42bfab31-c501-4a3d-9625-d95c2cb43ffa-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.574418 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42bfab31-c501-4a3d-9625-d95c2cb43ffa" (UID: "42bfab31-c501-4a3d-9625-d95c2cb43ffa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.604550 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.685160 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-config-data" (OuterVolumeSpecName: "config-data") pod "42bfab31-c501-4a3d-9625-d95c2cb43ffa" (UID: "42bfab31-c501-4a3d-9625-d95c2cb43ffa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.709987 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42bfab31-c501-4a3d-9625-d95c2cb43ffa-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.813726 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.866456 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e04bffff-693d-4909-914d-1e2f784ca47d" path="/var/lib/kubelet/pods/e04bffff-693d-4909-914d-1e2f784ca47d/volumes" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.908338 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-r8rpn"] Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.908585 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" podUID="bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" containerName="dnsmasq-dns" containerID="cri-o://e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463" gracePeriod=10 Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.956922 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7c64685c4b-rgnfp" event={"ID":"d9fd2d2f-01df-4aa4-bc83-46f30fb0c592","Type":"ContainerStarted","Data":"7f5132617bdcaf5f305ce9f399ca3c03d88f6af067a43e274339bb264abf72c8"} Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.957433 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.966154 4857 generic.go:334] "Generic (PLEG): container finished" podID="e8524783-2889-48f9-b8f9-0d5fb3db15e5" containerID="3ceef062cfc5b1739b6a96fda7934b07b46e681ebc746c7ee514e7d19eca5aeb" exitCode=1 Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.966873 4857 scope.go:117] "RemoveContainer" containerID="3ceef062cfc5b1739b6a96fda7934b07b46e681ebc746c7ee514e7d19eca5aeb" Dec 01 21:56:07 crc kubenswrapper[4857]: I1201 21:56:07.967164 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f87845c79-xkcmc" event={"ID":"e8524783-2889-48f9-b8f9-0d5fb3db15e5","Type":"ContainerDied","Data":"3ceef062cfc5b1739b6a96fda7934b07b46e681ebc746c7ee514e7d19eca5aeb"} Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.001559 4857 generic.go:334] "Generic (PLEG): container finished" podID="9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" containerID="01f58acb020e10d28a6a74e9effd07f47b94e4b7fdbcdef1da6efea79611af09" exitCode=1 Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.001632 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" event={"ID":"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24","Type":"ContainerDied","Data":"01f58acb020e10d28a6a74e9effd07f47b94e4b7fdbcdef1da6efea79611af09"} Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.002385 4857 scope.go:117] "RemoveContainer" containerID="01f58acb020e10d28a6a74e9effd07f47b94e4b7fdbcdef1da6efea79611af09" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.004883 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7c64685c4b-rgnfp" podStartSLOduration=4.004853308 podStartE2EDuration="4.004853308s" podCreationTimestamp="2025-12-01 21:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:07.986536039 +0000 UTC m=+1286.476598356" watchObservedRunningTime="2025-12-01 21:56:08.004853308 +0000 UTC m=+1286.494915625" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.063997 4857 generic.go:334] "Generic (PLEG): container finished" podID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerID="711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f" exitCode=0 Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.064998 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.065512 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42bfab31-c501-4a3d-9625-d95c2cb43ffa","Type":"ContainerDied","Data":"711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f"} Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.065552 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42bfab31-c501-4a3d-9625-d95c2cb43ffa","Type":"ContainerDied","Data":"d3e2e01c677a5acc95073859a6e571c0513325988d462a329a78d150ed84e96b"} Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.065570 4857 scope.go:117] "RemoveContainer" containerID="6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.100918 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.155163 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.183672 4857 scope.go:117] "RemoveContainer" containerID="f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.193505 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:56:08 crc kubenswrapper[4857]: E1201 21:56:08.193982 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="ceilometer-notification-agent" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.193994 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="ceilometer-notification-agent" Dec 01 21:56:08 crc kubenswrapper[4857]: E1201 21:56:08.194008 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="proxy-httpd" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.194015 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="proxy-httpd" Dec 01 21:56:08 crc kubenswrapper[4857]: E1201 21:56:08.194034 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="sg-core" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.194054 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="sg-core" Dec 01 21:56:08 crc kubenswrapper[4857]: E1201 21:56:08.194067 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e04bffff-693d-4909-914d-1e2f784ca47d" containerName="neutron-httpd" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.194073 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e04bffff-693d-4909-914d-1e2f784ca47d" containerName="neutron-httpd" Dec 01 21:56:08 crc kubenswrapper[4857]: E1201 21:56:08.194084 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e04bffff-693d-4909-914d-1e2f784ca47d" containerName="neutron-api" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.194089 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e04bffff-693d-4909-914d-1e2f784ca47d" containerName="neutron-api" Dec 01 21:56:08 crc kubenswrapper[4857]: E1201 21:56:08.194100 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="ceilometer-central-agent" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.194105 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="ceilometer-central-agent" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.194300 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="ceilometer-notification-agent" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.194314 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e04bffff-693d-4909-914d-1e2f784ca47d" containerName="neutron-httpd" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.194322 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="ceilometer-central-agent" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.194329 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e04bffff-693d-4909-914d-1e2f784ca47d" containerName="neutron-api" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.194342 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="proxy-httpd" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.194358 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" containerName="sg-core" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.196929 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.201078 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.201355 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.202201 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.225507 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-scripts\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.225573 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913cd3df-6038-44e3-ad22-fc0fa6645058-log-httpd\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.225597 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v586f\" (UniqueName: \"kubernetes.io/projected/913cd3df-6038-44e3-ad22-fc0fa6645058-kube-api-access-v586f\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.225651 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-config-data\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.225673 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.225704 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.225743 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913cd3df-6038-44e3-ad22-fc0fa6645058-run-httpd\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.236320 4857 scope.go:117] "RemoveContainer" containerID="711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.325281 4857 scope.go:117] "RemoveContainer" containerID="e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.327566 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913cd3df-6038-44e3-ad22-fc0fa6645058-run-httpd\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.327665 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-scripts\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.327707 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913cd3df-6038-44e3-ad22-fc0fa6645058-log-httpd\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.327729 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v586f\" (UniqueName: \"kubernetes.io/projected/913cd3df-6038-44e3-ad22-fc0fa6645058-kube-api-access-v586f\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.327777 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-config-data\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.327797 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.327828 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.331964 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913cd3df-6038-44e3-ad22-fc0fa6645058-log-httpd\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.332352 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913cd3df-6038-44e3-ad22-fc0fa6645058-run-httpd\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.341300 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.341489 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-scripts\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.342945 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.361157 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-77bf9494c9-4bdzk"] Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.361410 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-77bf9494c9-4bdzk" podUID="502899ed-eb9a-4eae-a3f3-1ef7369c08e7" containerName="heat-api" containerID="cri-o://b97342b2bd102c51865c503ee965467bd9493b0f742e791191d394191444f65b" gracePeriod=60 Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.363729 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-config-data\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.375312 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/heat-api-77bf9494c9-4bdzk" podUID="502899ed-eb9a-4eae-a3f3-1ef7369c08e7" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.169:8004/healthcheck\": EOF" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.383367 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-75f4dbf68b-t5t5b"] Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.383619 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" podUID="69afe439-86e6-490f-9476-ebb89b540192" containerName="heat-cfnapi" containerID="cri-o://5739d985f701aa8a38cd890c6e351dcaf126790bc2552d67e58b2f5f879831e2" gracePeriod=60 Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.396492 4857 scope.go:117] "RemoveContainer" containerID="6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db" Dec 01 21:56:08 crc kubenswrapper[4857]: E1201 21:56:08.404460 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db\": container with ID starting with 6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db not found: ID does not exist" containerID="6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.404512 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db"} err="failed to get container status \"6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db\": rpc error: code = NotFound desc = could not find container \"6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db\": container with ID starting with 6a7fbd64ddead0d43a7bde578d98922cd8bcf768664c4d935fb4d565832b36db not found: ID does not exist" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.404537 4857 scope.go:117] "RemoveContainer" containerID="f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4" Dec 01 21:56:08 crc kubenswrapper[4857]: E1201 21:56:08.404776 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4\": container with ID starting with f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4 not found: ID does not exist" containerID="f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.404798 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4"} err="failed to get container status \"f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4\": rpc error: code = NotFound desc = could not find container \"f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4\": container with ID starting with f2fb6ff3a44712636a5df78df7a43b35e77bb68827957587966cd3220eb512d4 not found: ID does not exist" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.404811 4857 scope.go:117] "RemoveContainer" containerID="711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f" Dec 01 21:56:08 crc kubenswrapper[4857]: E1201 21:56:08.407232 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f\": container with ID starting with 711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f not found: ID does not exist" containerID="711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.407254 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f"} err="failed to get container status \"711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f\": rpc error: code = NotFound desc = could not find container \"711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f\": container with ID starting with 711c6641ebacb8883e222c3f539610c9a925ca8479d3782852b47f8491426f2f not found: ID does not exist" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.407269 4857 scope.go:117] "RemoveContainer" containerID="e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.409664 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v586f\" (UniqueName: \"kubernetes.io/projected/913cd3df-6038-44e3-ad22-fc0fa6645058-kube-api-access-v586f\") pod \"ceilometer-0\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: E1201 21:56:08.413364 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2\": container with ID starting with e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2 not found: ID does not exist" containerID="e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.413402 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2"} err="failed to get container status \"e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2\": rpc error: code = NotFound desc = could not find container \"e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2\": container with ID starting with e9eb330c21b52e212db2dd881f1c144e42b227cb37cbf2f7419f58d5b47d17f2 not found: ID does not exist" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.493983 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5c86f5f679-g6dgl"] Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.495423 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.507186 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-77bf9494c9-4bdzk" podUID="502899ed-eb9a-4eae-a3f3-1ef7369c08e7" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.169:8004/healthcheck\": EOF" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.511031 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.511836 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.515927 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.521453 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5c86f5f679-g6dgl"] Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.532526 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-internal-tls-certs\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.532578 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-combined-ca-bundle\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.532655 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-config-data\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.532673 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-config-data-custom\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.532712 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hjj2\" (UniqueName: \"kubernetes.io/projected/05d88f39-b756-4b31-9385-99213e839003-kube-api-access-6hjj2\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.532729 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-public-tls-certs\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.560192 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5894b69764-fpz5r"] Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.561902 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.577137 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.577306 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.629074 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5894b69764-fpz5r"] Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.635218 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-config-data-custom\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.635266 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-config-data\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.635301 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-config-data\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.635331 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7td6n\" (UniqueName: \"kubernetes.io/projected/ce4b4968-ffc0-4d0d-86b5-da72a0712630-kube-api-access-7td6n\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.635350 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hjj2\" (UniqueName: \"kubernetes.io/projected/05d88f39-b756-4b31-9385-99213e839003-kube-api-access-6hjj2\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.635369 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-public-tls-certs\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.635405 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-internal-tls-certs\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.635425 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-config-data-custom\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.635466 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-public-tls-certs\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.635484 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-internal-tls-certs\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.635501 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-combined-ca-bundle\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.635519 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-combined-ca-bundle\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.646976 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-combined-ca-bundle\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.666748 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-public-tls-certs\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.675719 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-internal-tls-certs\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.682382 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-config-data-custom\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.682412 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05d88f39-b756-4b31-9385-99213e839003-config-data\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.682754 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hjj2\" (UniqueName: \"kubernetes.io/projected/05d88f39-b756-4b31-9385-99213e839003-kube-api-access-6hjj2\") pod \"heat-api-5c86f5f679-g6dgl\" (UID: \"05d88f39-b756-4b31-9385-99213e839003\") " pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.686134 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" podUID="69afe439-86e6-490f-9476-ebb89b540192" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.167:8000/healthcheck\": EOF" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.760480 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-internal-tls-certs\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.760550 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-config-data-custom\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.760615 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-public-tls-certs\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.760650 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-combined-ca-bundle\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.760775 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-config-data\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.760814 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7td6n\" (UniqueName: \"kubernetes.io/projected/ce4b4968-ffc0-4d0d-86b5-da72a0712630-kube-api-access-7td6n\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.792070 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-public-tls-certs\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.792930 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-config-data-custom\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.804754 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-config-data\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.811095 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-internal-tls-certs\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.811913 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce4b4968-ffc0-4d0d-86b5-da72a0712630-combined-ca-bundle\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.815955 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7td6n\" (UniqueName: \"kubernetes.io/projected/ce4b4968-ffc0-4d0d-86b5-da72a0712630-kube-api-access-7td6n\") pod \"heat-cfnapi-5894b69764-fpz5r\" (UID: \"ce4b4968-ffc0-4d0d-86b5-da72a0712630\") " pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.824720 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.850050 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.863331 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.965628 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-ovsdbserver-sb\") pod \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.965742 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c59jz\" (UniqueName: \"kubernetes.io/projected/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-kube-api-access-c59jz\") pod \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.965766 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-dns-swift-storage-0\") pod \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.965828 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-ovsdbserver-nb\") pod \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.965865 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-config\") pod \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " Dec 01 21:56:08 crc kubenswrapper[4857]: I1201 21:56:08.966019 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-dns-svc\") pod \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\" (UID: \"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12\") " Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.013194 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-kube-api-access-c59jz" (OuterVolumeSpecName: "kube-api-access-c59jz") pod "bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" (UID: "bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12"). InnerVolumeSpecName "kube-api-access-c59jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.050875 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" (UID: "bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.069345 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.069398 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c59jz\" (UniqueName: \"kubernetes.io/projected/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-kube-api-access-c59jz\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.099602 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" (UID: "bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.106768 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-config" (OuterVolumeSpecName: "config") pod "bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" (UID: "bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.136855 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" (UID: "bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.141521 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" (UID: "bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.150214 4857 generic.go:334] "Generic (PLEG): container finished" podID="bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" containerID="e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463" exitCode=0 Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.150265 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" event={"ID":"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12","Type":"ContainerDied","Data":"e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463"} Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.150294 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" event={"ID":"bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12","Type":"ContainerDied","Data":"f37419770688b54e6d2c6c2f0df62d40fbfee22c53d9fbd59923f4e913d3dae8"} Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.150311 4857 scope.go:117] "RemoveContainer" containerID="e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.150410 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.170570 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f87845c79-xkcmc" event={"ID":"e8524783-2889-48f9-b8f9-0d5fb3db15e5","Type":"ContainerStarted","Data":"bd399d9bc3979936cbf901415e37c64434e266d2e56e67201d609de9a207d0b6"} Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.170818 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.170921 4857 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.170952 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.170963 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.170971 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.177614 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" event={"ID":"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24","Type":"ContainerStarted","Data":"e63cb4ca5c7b1ea70185621cc0ce903fcf6418219b29fd3166b264329ce01a47"} Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.177683 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.198879 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-f87845c79-xkcmc" podStartSLOduration=5.198862054 podStartE2EDuration="5.198862054s" podCreationTimestamp="2025-12-01 21:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:09.189588777 +0000 UTC m=+1287.679651094" watchObservedRunningTime="2025-12-01 21:56:09.198862054 +0000 UTC m=+1287.688924371" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.225131 4857 scope.go:117] "RemoveContainer" containerID="9c57cf46fd8ad499bf7a40af7b25958254da9960c76192fc72491ff863ece35d" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.233911 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-r8rpn"] Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.263108 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-r8rpn"] Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.267049 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" podStartSLOduration=5.267018415 podStartE2EDuration="5.267018415s" podCreationTimestamp="2025-12-01 21:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:09.227381473 +0000 UTC m=+1287.717443790" watchObservedRunningTime="2025-12-01 21:56:09.267018415 +0000 UTC m=+1287.757080732" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.292275 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.292508 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="07e01fec-a494-44c1-9671-b5f32f79e96a" containerName="glance-log" containerID="cri-o://4925b165c7077377e62bea2da2c5c31c88ce758e3a6407de1e148ddf0131b679" gracePeriod=30 Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.292952 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="07e01fec-a494-44c1-9671-b5f32f79e96a" containerName="glance-httpd" containerID="cri-o://52086b2db0db0092223027f370cb55c8d6cbdd2a7c93d9296553b1a04101e5db" gracePeriod=30 Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.297034 4857 scope.go:117] "RemoveContainer" containerID="e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463" Dec 01 21:56:09 crc kubenswrapper[4857]: E1201 21:56:09.302434 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463\": container with ID starting with e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463 not found: ID does not exist" containerID="e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.302476 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463"} err="failed to get container status \"e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463\": rpc error: code = NotFound desc = could not find container \"e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463\": container with ID starting with e15b3a5d19377827de61d6fe885e144eebbde2152c3bb844cb85ca6b4feca463 not found: ID does not exist" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.302499 4857 scope.go:117] "RemoveContainer" containerID="9c57cf46fd8ad499bf7a40af7b25958254da9960c76192fc72491ff863ece35d" Dec 01 21:56:09 crc kubenswrapper[4857]: E1201 21:56:09.304196 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c57cf46fd8ad499bf7a40af7b25958254da9960c76192fc72491ff863ece35d\": container with ID starting with 9c57cf46fd8ad499bf7a40af7b25958254da9960c76192fc72491ff863ece35d not found: ID does not exist" containerID="9c57cf46fd8ad499bf7a40af7b25958254da9960c76192fc72491ff863ece35d" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.304236 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c57cf46fd8ad499bf7a40af7b25958254da9960c76192fc72491ff863ece35d"} err="failed to get container status \"9c57cf46fd8ad499bf7a40af7b25958254da9960c76192fc72491ff863ece35d\": rpc error: code = NotFound desc = could not find container \"9c57cf46fd8ad499bf7a40af7b25958254da9960c76192fc72491ff863ece35d\": container with ID starting with 9c57cf46fd8ad499bf7a40af7b25958254da9960c76192fc72491ff863ece35d not found: ID does not exist" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.315310 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.493113 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5c86f5f679-g6dgl"] Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.669841 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5894b69764-fpz5r"] Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.872315 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42bfab31-c501-4a3d-9625-d95c2cb43ffa" path="/var/lib/kubelet/pods/42bfab31-c501-4a3d-9625-d95c2cb43ffa/volumes" Dec 01 21:56:09 crc kubenswrapper[4857]: I1201 21:56:09.873099 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" path="/var/lib/kubelet/pods/bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12/volumes" Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.169402 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.202644 4857 generic.go:334] "Generic (PLEG): container finished" podID="e8524783-2889-48f9-b8f9-0d5fb3db15e5" containerID="bd399d9bc3979936cbf901415e37c64434e266d2e56e67201d609de9a207d0b6" exitCode=1 Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.202744 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f87845c79-xkcmc" event={"ID":"e8524783-2889-48f9-b8f9-0d5fb3db15e5","Type":"ContainerDied","Data":"bd399d9bc3979936cbf901415e37c64434e266d2e56e67201d609de9a207d0b6"} Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.202783 4857 scope.go:117] "RemoveContainer" containerID="3ceef062cfc5b1739b6a96fda7934b07b46e681ebc746c7ee514e7d19eca5aeb" Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.203365 4857 scope.go:117] "RemoveContainer" containerID="bd399d9bc3979936cbf901415e37c64434e266d2e56e67201d609de9a207d0b6" Dec 01 21:56:10 crc kubenswrapper[4857]: E1201 21:56:10.203687 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-f87845c79-xkcmc_openstack(e8524783-2889-48f9-b8f9-0d5fb3db15e5)\"" pod="openstack/heat-api-f87845c79-xkcmc" podUID="e8524783-2889-48f9-b8f9-0d5fb3db15e5" Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.207478 4857 generic.go:334] "Generic (PLEG): container finished" podID="07e01fec-a494-44c1-9671-b5f32f79e96a" containerID="4925b165c7077377e62bea2da2c5c31c88ce758e3a6407de1e148ddf0131b679" exitCode=143 Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.207551 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"07e01fec-a494-44c1-9671-b5f32f79e96a","Type":"ContainerDied","Data":"4925b165c7077377e62bea2da2c5c31c88ce758e3a6407de1e148ddf0131b679"} Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.215019 4857 generic.go:334] "Generic (PLEG): container finished" podID="9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" containerID="e63cb4ca5c7b1ea70185621cc0ce903fcf6418219b29fd3166b264329ce01a47" exitCode=1 Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.215102 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" event={"ID":"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24","Type":"ContainerDied","Data":"e63cb4ca5c7b1ea70185621cc0ce903fcf6418219b29fd3166b264329ce01a47"} Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.216004 4857 scope.go:117] "RemoveContainer" containerID="e63cb4ca5c7b1ea70185621cc0ce903fcf6418219b29fd3166b264329ce01a47" Dec 01 21:56:10 crc kubenswrapper[4857]: E1201 21:56:10.216345 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-66cd8df7d6-zlb8c_openstack(9dff5cc8-16db-4b2b-8d0a-72eb2c068c24)\"" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" podUID="9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.229191 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5c86f5f679-g6dgl" event={"ID":"05d88f39-b756-4b31-9385-99213e839003","Type":"ContainerStarted","Data":"29d1d4ccea17e9eb154ba7c7ece911b0802f2210af65e0f0857858d97382f547"} Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.229236 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5c86f5f679-g6dgl" event={"ID":"05d88f39-b756-4b31-9385-99213e839003","Type":"ContainerStarted","Data":"d8f3060a0aec2d50d27ef2c3af044ab38fd1af5ed6e034b8063b3675ece312e4"} Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.230007 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.231919 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5894b69764-fpz5r" event={"ID":"ce4b4968-ffc0-4d0d-86b5-da72a0712630","Type":"ContainerStarted","Data":"a3b7d504802a8c87d01cca865e57783567b3262fdcfa6f61e1e2f2f761edbcb4"} Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.231957 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5894b69764-fpz5r" event={"ID":"ce4b4968-ffc0-4d0d-86b5-da72a0712630","Type":"ContainerStarted","Data":"1478b09279c96c87e359688f52552730f33145cf5dd32bde0ba27a3bfeb3af40"} Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.232171 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.246256 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913cd3df-6038-44e3-ad22-fc0fa6645058","Type":"ContainerStarted","Data":"04c300b0475348f55f4b71cccecccf3b2287989e898aab789a62953827605f01"} Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.261145 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.267570 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.275744 4857 scope.go:117] "RemoveContainer" containerID="01f58acb020e10d28a6a74e9effd07f47b94e4b7fdbcdef1da6efea79611af09" Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.296979 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5c86f5f679-g6dgl" podStartSLOduration=2.29696074 podStartE2EDuration="2.29696074s" podCreationTimestamp="2025-12-01 21:56:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:10.28716793 +0000 UTC m=+1288.777230267" watchObservedRunningTime="2025-12-01 21:56:10.29696074 +0000 UTC m=+1288.787023057" Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.322460 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5894b69764-fpz5r" podStartSLOduration=2.322441015 podStartE2EDuration="2.322441015s" podCreationTimestamp="2025-12-01 21:56:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:10.313398073 +0000 UTC m=+1288.803460390" watchObservedRunningTime="2025-12-01 21:56:10.322441015 +0000 UTC m=+1288.812503332" Dec 01 21:56:10 crc kubenswrapper[4857]: I1201 21:56:10.563201 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 21:56:11 crc kubenswrapper[4857]: I1201 21:56:11.259810 4857 scope.go:117] "RemoveContainer" containerID="bd399d9bc3979936cbf901415e37c64434e266d2e56e67201d609de9a207d0b6" Dec 01 21:56:11 crc kubenswrapper[4857]: I1201 21:56:11.260199 4857 scope.go:117] "RemoveContainer" containerID="e63cb4ca5c7b1ea70185621cc0ce903fcf6418219b29fd3166b264329ce01a47" Dec 01 21:56:11 crc kubenswrapper[4857]: E1201 21:56:11.260465 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-66cd8df7d6-zlb8c_openstack(9dff5cc8-16db-4b2b-8d0a-72eb2c068c24)\"" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" podUID="9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" Dec 01 21:56:11 crc kubenswrapper[4857]: E1201 21:56:11.260816 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-f87845c79-xkcmc_openstack(e8524783-2889-48f9-b8f9-0d5fb3db15e5)\"" pod="openstack/heat-api-f87845c79-xkcmc" podUID="e8524783-2889-48f9-b8f9-0d5fb3db15e5" Dec 01 21:56:11 crc kubenswrapper[4857]: I1201 21:56:11.268295 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913cd3df-6038-44e3-ad22-fc0fa6645058","Type":"ContainerStarted","Data":"d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1"} Dec 01 21:56:11 crc kubenswrapper[4857]: I1201 21:56:11.954274 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:11 crc kubenswrapper[4857]: I1201 21:56:11.956310 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b8476f857-zrr4q" Dec 01 21:56:12 crc kubenswrapper[4857]: I1201 21:56:12.415568 4857 scope.go:117] "RemoveContainer" containerID="e63cb4ca5c7b1ea70185621cc0ce903fcf6418219b29fd3166b264329ce01a47" Dec 01 21:56:12 crc kubenswrapper[4857]: I1201 21:56:12.416179 4857 scope.go:117] "RemoveContainer" containerID="bd399d9bc3979936cbf901415e37c64434e266d2e56e67201d609de9a207d0b6" Dec 01 21:56:12 crc kubenswrapper[4857]: E1201 21:56:12.416598 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-f87845c79-xkcmc_openstack(e8524783-2889-48f9-b8f9-0d5fb3db15e5)\"" pod="openstack/heat-api-f87845c79-xkcmc" podUID="e8524783-2889-48f9-b8f9-0d5fb3db15e5" Dec 01 21:56:12 crc kubenswrapper[4857]: E1201 21:56:12.418644 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-66cd8df7d6-zlb8c_openstack(9dff5cc8-16db-4b2b-8d0a-72eb2c068c24)\"" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" podUID="9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" Dec 01 21:56:13 crc kubenswrapper[4857]: I1201 21:56:13.426896 4857 generic.go:334] "Generic (PLEG): container finished" podID="07e01fec-a494-44c1-9671-b5f32f79e96a" containerID="52086b2db0db0092223027f370cb55c8d6cbdd2a7c93d9296553b1a04101e5db" exitCode=0 Dec 01 21:56:13 crc kubenswrapper[4857]: I1201 21:56:13.426967 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"07e01fec-a494-44c1-9671-b5f32f79e96a","Type":"ContainerDied","Data":"52086b2db0db0092223027f370cb55c8d6cbdd2a7c93d9296553b1a04101e5db"} Dec 01 21:56:13 crc kubenswrapper[4857]: I1201 21:56:13.596191 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c9776ccc5-r8rpn" podUID="bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.162:5353: i/o timeout" Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.830673 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-ktkxd"] Dec 01 21:56:15 crc kubenswrapper[4857]: E1201 21:56:15.832063 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" containerName="dnsmasq-dns" Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.832079 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" containerName="dnsmasq-dns" Dec 01 21:56:15 crc kubenswrapper[4857]: E1201 21:56:15.832102 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" containerName="init" Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.832109 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" containerName="init" Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.832692 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcac39ab-c5eb-4a8c-b313-b44bdc7c4a12" containerName="dnsmasq-dns" Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.833786 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ktkxd" Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.839752 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-77bf9494c9-4bdzk" podUID="502899ed-eb9a-4eae-a3f3-1ef7369c08e7" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.169:8004/healthcheck\": read tcp 10.217.0.2:47016->10.217.0.169:8004: read: connection reset by peer" Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.840508 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-77bf9494c9-4bdzk" podUID="502899ed-eb9a-4eae-a3f3-1ef7369c08e7" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.169:8004/healthcheck\": dial tcp 10.217.0.169:8004: connect: connection refused" Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.861777 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ktkxd"] Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.879790 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtww2\" (UniqueName: \"kubernetes.io/projected/cc85532c-b753-4a79-ad36-621f558d03d0-kube-api-access-vtww2\") pod \"nova-api-db-create-ktkxd\" (UID: \"cc85532c-b753-4a79-ad36-621f558d03d0\") " pod="openstack/nova-api-db-create-ktkxd" Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.880108 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc85532c-b753-4a79-ad36-621f558d03d0-operator-scripts\") pod \"nova-api-db-create-ktkxd\" (UID: \"cc85532c-b753-4a79-ad36-621f558d03d0\") " pod="openstack/nova-api-db-create-ktkxd" Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.986278 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtww2\" (UniqueName: \"kubernetes.io/projected/cc85532c-b753-4a79-ad36-621f558d03d0-kube-api-access-vtww2\") pod \"nova-api-db-create-ktkxd\" (UID: \"cc85532c-b753-4a79-ad36-621f558d03d0\") " pod="openstack/nova-api-db-create-ktkxd" Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.986417 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc85532c-b753-4a79-ad36-621f558d03d0-operator-scripts\") pod \"nova-api-db-create-ktkxd\" (UID: \"cc85532c-b753-4a79-ad36-621f558d03d0\") " pod="openstack/nova-api-db-create-ktkxd" Dec 01 21:56:15 crc kubenswrapper[4857]: I1201 21:56:15.987350 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc85532c-b753-4a79-ad36-621f558d03d0-operator-scripts\") pod \"nova-api-db-create-ktkxd\" (UID: \"cc85532c-b753-4a79-ad36-621f558d03d0\") " pod="openstack/nova-api-db-create-ktkxd" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.045851 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-cjk48"] Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.046268 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtww2\" (UniqueName: \"kubernetes.io/projected/cc85532c-b753-4a79-ad36-621f558d03d0-kube-api-access-vtww2\") pod \"nova-api-db-create-ktkxd\" (UID: \"cc85532c-b753-4a79-ad36-621f558d03d0\") " pod="openstack/nova-api-db-create-ktkxd" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.047386 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cjk48" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.089767 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cjk48"] Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.095075 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9e73cb1-7588-46b5-9014-efe779354e1d-operator-scripts\") pod \"nova-cell0-db-create-cjk48\" (UID: \"d9e73cb1-7588-46b5-9014-efe779354e1d\") " pod="openstack/nova-cell0-db-create-cjk48" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.095369 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxgcw\" (UniqueName: \"kubernetes.io/projected/d9e73cb1-7588-46b5-9014-efe779354e1d-kube-api-access-nxgcw\") pod \"nova-cell0-db-create-cjk48\" (UID: \"d9e73cb1-7588-46b5-9014-efe779354e1d\") " pod="openstack/nova-cell0-db-create-cjk48" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.158821 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-dhc4w"] Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.161580 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dhc4w" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.169669 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dhc4w"] Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.193639 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-cb18-account-create-update-q2fbk"] Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.197473 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-cb18-account-create-update-q2fbk" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.198720 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxgcw\" (UniqueName: \"kubernetes.io/projected/d9e73cb1-7588-46b5-9014-efe779354e1d-kube-api-access-nxgcw\") pod \"nova-cell0-db-create-cjk48\" (UID: \"d9e73cb1-7588-46b5-9014-efe779354e1d\") " pod="openstack/nova-cell0-db-create-cjk48" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.198878 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmqs8\" (UniqueName: \"kubernetes.io/projected/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c-kube-api-access-jmqs8\") pod \"nova-cell1-db-create-dhc4w\" (UID: \"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c\") " pod="openstack/nova-cell1-db-create-dhc4w" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.198935 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c-operator-scripts\") pod \"nova-cell1-db-create-dhc4w\" (UID: \"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c\") " pod="openstack/nova-cell1-db-create-dhc4w" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.199028 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9e73cb1-7588-46b5-9014-efe779354e1d-operator-scripts\") pod \"nova-cell0-db-create-cjk48\" (UID: \"d9e73cb1-7588-46b5-9014-efe779354e1d\") " pod="openstack/nova-cell0-db-create-cjk48" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.201678 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.206571 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9e73cb1-7588-46b5-9014-efe779354e1d-operator-scripts\") pod \"nova-cell0-db-create-cjk48\" (UID: \"d9e73cb1-7588-46b5-9014-efe779354e1d\") " pod="openstack/nova-cell0-db-create-cjk48" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.219603 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-cb18-account-create-update-q2fbk"] Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.247766 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxgcw\" (UniqueName: \"kubernetes.io/projected/d9e73cb1-7588-46b5-9014-efe779354e1d-kube-api-access-nxgcw\") pod \"nova-cell0-db-create-cjk48\" (UID: \"d9e73cb1-7588-46b5-9014-efe779354e1d\") " pod="openstack/nova-cell0-db-create-cjk48" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.269126 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ktkxd" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.304656 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7174f1f-8255-4dfe-a230-5128ba63571e-operator-scripts\") pod \"nova-api-cb18-account-create-update-q2fbk\" (UID: \"a7174f1f-8255-4dfe-a230-5128ba63571e\") " pod="openstack/nova-api-cb18-account-create-update-q2fbk" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.304783 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttr4d\" (UniqueName: \"kubernetes.io/projected/a7174f1f-8255-4dfe-a230-5128ba63571e-kube-api-access-ttr4d\") pod \"nova-api-cb18-account-create-update-q2fbk\" (UID: \"a7174f1f-8255-4dfe-a230-5128ba63571e\") " pod="openstack/nova-api-cb18-account-create-update-q2fbk" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.304843 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmqs8\" (UniqueName: \"kubernetes.io/projected/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c-kube-api-access-jmqs8\") pod \"nova-cell1-db-create-dhc4w\" (UID: \"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c\") " pod="openstack/nova-cell1-db-create-dhc4w" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.304892 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c-operator-scripts\") pod \"nova-cell1-db-create-dhc4w\" (UID: \"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c\") " pod="openstack/nova-cell1-db-create-dhc4w" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.306139 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c-operator-scripts\") pod \"nova-cell1-db-create-dhc4w\" (UID: \"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c\") " pod="openstack/nova-cell1-db-create-dhc4w" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.339396 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmqs8\" (UniqueName: \"kubernetes.io/projected/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c-kube-api-access-jmqs8\") pod \"nova-cell1-db-create-dhc4w\" (UID: \"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c\") " pod="openstack/nova-cell1-db-create-dhc4w" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.341462 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-7c37-account-create-update-rzsv8"] Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.343021 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.346303 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.350401 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7c37-account-create-update-rzsv8"] Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.362711 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-d8e3-account-create-update-nrctx"] Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.364917 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.371523 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.406574 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d8e3-account-create-update-nrctx"] Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.407684 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9638968d-42b2-46ab-b02a-72d044325fa9-operator-scripts\") pod \"nova-cell1-d8e3-account-create-update-nrctx\" (UID: \"9638968d-42b2-46ab-b02a-72d044325fa9\") " pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.407907 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bw6j\" (UniqueName: \"kubernetes.io/projected/9638968d-42b2-46ab-b02a-72d044325fa9-kube-api-access-8bw6j\") pod \"nova-cell1-d8e3-account-create-update-nrctx\" (UID: \"9638968d-42b2-46ab-b02a-72d044325fa9\") " pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.408095 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7174f1f-8255-4dfe-a230-5128ba63571e-operator-scripts\") pod \"nova-api-cb18-account-create-update-q2fbk\" (UID: \"a7174f1f-8255-4dfe-a230-5128ba63571e\") " pod="openstack/nova-api-cb18-account-create-update-q2fbk" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.408257 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttr4d\" (UniqueName: \"kubernetes.io/projected/a7174f1f-8255-4dfe-a230-5128ba63571e-kube-api-access-ttr4d\") pod \"nova-api-cb18-account-create-update-q2fbk\" (UID: \"a7174f1f-8255-4dfe-a230-5128ba63571e\") " pod="openstack/nova-api-cb18-account-create-update-q2fbk" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.408344 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/574a17c8-2c54-41d2-8390-aff1f2220638-operator-scripts\") pod \"nova-cell0-7c37-account-create-update-rzsv8\" (UID: \"574a17c8-2c54-41d2-8390-aff1f2220638\") " pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.408712 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwswz\" (UniqueName: \"kubernetes.io/projected/574a17c8-2c54-41d2-8390-aff1f2220638-kube-api-access-dwswz\") pod \"nova-cell0-7c37-account-create-update-rzsv8\" (UID: \"574a17c8-2c54-41d2-8390-aff1f2220638\") " pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.409828 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" podUID="69afe439-86e6-490f-9476-ebb89b540192" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.167:8000/healthcheck\": read tcp 10.217.0.2:50858->10.217.0.167:8000: read: connection reset by peer" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.412584 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7174f1f-8255-4dfe-a230-5128ba63571e-operator-scripts\") pod \"nova-api-cb18-account-create-update-q2fbk\" (UID: \"a7174f1f-8255-4dfe-a230-5128ba63571e\") " pod="openstack/nova-api-cb18-account-create-update-q2fbk" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.438496 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttr4d\" (UniqueName: \"kubernetes.io/projected/a7174f1f-8255-4dfe-a230-5128ba63571e-kube-api-access-ttr4d\") pod \"nova-api-cb18-account-create-update-q2fbk\" (UID: \"a7174f1f-8255-4dfe-a230-5128ba63571e\") " pod="openstack/nova-api-cb18-account-create-update-q2fbk" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.443310 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cjk48" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.473879 4857 generic.go:334] "Generic (PLEG): container finished" podID="502899ed-eb9a-4eae-a3f3-1ef7369c08e7" containerID="b97342b2bd102c51865c503ee965467bd9493b0f742e791191d394191444f65b" exitCode=0 Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.473932 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77bf9494c9-4bdzk" event={"ID":"502899ed-eb9a-4eae-a3f3-1ef7369c08e7","Type":"ContainerDied","Data":"b97342b2bd102c51865c503ee965467bd9493b0f742e791191d394191444f65b"} Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.492405 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dhc4w" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.510483 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/574a17c8-2c54-41d2-8390-aff1f2220638-operator-scripts\") pod \"nova-cell0-7c37-account-create-update-rzsv8\" (UID: \"574a17c8-2c54-41d2-8390-aff1f2220638\") " pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.510551 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwswz\" (UniqueName: \"kubernetes.io/projected/574a17c8-2c54-41d2-8390-aff1f2220638-kube-api-access-dwswz\") pod \"nova-cell0-7c37-account-create-update-rzsv8\" (UID: \"574a17c8-2c54-41d2-8390-aff1f2220638\") " pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.510588 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9638968d-42b2-46ab-b02a-72d044325fa9-operator-scripts\") pod \"nova-cell1-d8e3-account-create-update-nrctx\" (UID: \"9638968d-42b2-46ab-b02a-72d044325fa9\") " pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.510644 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bw6j\" (UniqueName: \"kubernetes.io/projected/9638968d-42b2-46ab-b02a-72d044325fa9-kube-api-access-8bw6j\") pod \"nova-cell1-d8e3-account-create-update-nrctx\" (UID: \"9638968d-42b2-46ab-b02a-72d044325fa9\") " pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.512243 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/574a17c8-2c54-41d2-8390-aff1f2220638-operator-scripts\") pod \"nova-cell0-7c37-account-create-update-rzsv8\" (UID: \"574a17c8-2c54-41d2-8390-aff1f2220638\") " pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.513089 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9638968d-42b2-46ab-b02a-72d044325fa9-operator-scripts\") pod \"nova-cell1-d8e3-account-create-update-nrctx\" (UID: \"9638968d-42b2-46ab-b02a-72d044325fa9\") " pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.528076 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwswz\" (UniqueName: \"kubernetes.io/projected/574a17c8-2c54-41d2-8390-aff1f2220638-kube-api-access-dwswz\") pod \"nova-cell0-7c37-account-create-update-rzsv8\" (UID: \"574a17c8-2c54-41d2-8390-aff1f2220638\") " pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.529090 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bw6j\" (UniqueName: \"kubernetes.io/projected/9638968d-42b2-46ab-b02a-72d044325fa9-kube-api-access-8bw6j\") pod \"nova-cell1-d8e3-account-create-update-nrctx\" (UID: \"9638968d-42b2-46ab-b02a-72d044325fa9\") " pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.582014 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-cb18-account-create-update-q2fbk" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.598403 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.645133 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:56:16 crc kubenswrapper[4857]: I1201 21:56:16.738461 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" Dec 01 21:56:17 crc kubenswrapper[4857]: I1201 21:56:17.507897 4857 generic.go:334] "Generic (PLEG): container finished" podID="69afe439-86e6-490f-9476-ebb89b540192" containerID="5739d985f701aa8a38cd890c6e351dcaf126790bc2552d67e58b2f5f879831e2" exitCode=0 Dec 01 21:56:17 crc kubenswrapper[4857]: I1201 21:56:17.507997 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" event={"ID":"69afe439-86e6-490f-9476-ebb89b540192","Type":"ContainerDied","Data":"5739d985f701aa8a38cd890c6e351dcaf126790bc2552d67e58b2f5f879831e2"} Dec 01 21:56:17 crc kubenswrapper[4857]: I1201 21:56:17.740856 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:56:17 crc kubenswrapper[4857]: I1201 21:56:17.916134 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" podUID="69afe439-86e6-490f-9476-ebb89b540192" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.167:8000/healthcheck\": dial tcp 10.217.0.167:8000: connect: connection refused" Dec 01 21:56:18 crc kubenswrapper[4857]: I1201 21:56:18.143739 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-77bf9494c9-4bdzk" podUID="502899ed-eb9a-4eae-a3f3-1ef7369c08e7" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.169:8004/healthcheck\": dial tcp 10.217.0.169:8004: connect: connection refused" Dec 01 21:56:18 crc kubenswrapper[4857]: I1201 21:56:18.532787 4857 generic.go:334] "Generic (PLEG): container finished" podID="da1a9dd7-58b2-458e-8134-8251beec7d02" containerID="b0e084fffedc0dcf7b68c91f7c8bcbbed5b83dcbcbb7b5f088756c5a4cbbd871" exitCode=137 Dec 01 21:56:18 crc kubenswrapper[4857]: I1201 21:56:18.532877 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da1a9dd7-58b2-458e-8134-8251beec7d02","Type":"ContainerDied","Data":"b0e084fffedc0dcf7b68c91f7c8bcbbed5b83dcbcbb7b5f088756c5a4cbbd871"} Dec 01 21:56:18 crc kubenswrapper[4857]: I1201 21:56:18.711795 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="da1a9dd7-58b2-458e-8134-8251beec7d02" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.163:8776/healthcheck\": dial tcp 10.217.0.163:8776: connect: connection refused" Dec 01 21:56:19 crc kubenswrapper[4857]: I1201 21:56:19.093368 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:56:19 crc kubenswrapper[4857]: I1201 21:56:19.093751 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3ec1f836-90ab-4f44-b500-5a705a165925" containerName="glance-log" containerID="cri-o://152968eff1600d4b0bb4c27e45c9f98e2a9426bf72c50fed47e7d7d5af868bf2" gracePeriod=30 Dec 01 21:56:19 crc kubenswrapper[4857]: I1201 21:56:19.093887 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3ec1f836-90ab-4f44-b500-5a705a165925" containerName="glance-httpd" containerID="cri-o://0761be5c409a673752c97c6527eca4e0dabc3ce5b2e9374fa9db1ff742f67ccd" gracePeriod=30 Dec 01 21:56:19 crc kubenswrapper[4857]: I1201 21:56:19.544325 4857 generic.go:334] "Generic (PLEG): container finished" podID="3ec1f836-90ab-4f44-b500-5a705a165925" containerID="152968eff1600d4b0bb4c27e45c9f98e2a9426bf72c50fed47e7d7d5af868bf2" exitCode=143 Dec 01 21:56:19 crc kubenswrapper[4857]: I1201 21:56:19.544371 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3ec1f836-90ab-4f44-b500-5a705a165925","Type":"ContainerDied","Data":"152968eff1600d4b0bb4c27e45c9f98e2a9426bf72c50fed47e7d7d5af868bf2"} Dec 01 21:56:20 crc kubenswrapper[4857]: I1201 21:56:20.817203 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 21:56:20 crc kubenswrapper[4857]: I1201 21:56:20.968717 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpd85\" (UniqueName: \"kubernetes.io/projected/da1a9dd7-58b2-458e-8134-8251beec7d02-kube-api-access-vpd85\") pod \"da1a9dd7-58b2-458e-8134-8251beec7d02\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " Dec 01 21:56:20 crc kubenswrapper[4857]: I1201 21:56:20.969209 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-combined-ca-bundle\") pod \"da1a9dd7-58b2-458e-8134-8251beec7d02\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " Dec 01 21:56:20 crc kubenswrapper[4857]: I1201 21:56:20.969240 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da1a9dd7-58b2-458e-8134-8251beec7d02-logs\") pod \"da1a9dd7-58b2-458e-8134-8251beec7d02\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " Dec 01 21:56:20 crc kubenswrapper[4857]: I1201 21:56:20.969369 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-config-data-custom\") pod \"da1a9dd7-58b2-458e-8134-8251beec7d02\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " Dec 01 21:56:20 crc kubenswrapper[4857]: I1201 21:56:20.969414 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-scripts\") pod \"da1a9dd7-58b2-458e-8134-8251beec7d02\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " Dec 01 21:56:20 crc kubenswrapper[4857]: I1201 21:56:20.969498 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da1a9dd7-58b2-458e-8134-8251beec7d02-etc-machine-id\") pod \"da1a9dd7-58b2-458e-8134-8251beec7d02\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " Dec 01 21:56:20 crc kubenswrapper[4857]: I1201 21:56:20.969523 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-config-data\") pod \"da1a9dd7-58b2-458e-8134-8251beec7d02\" (UID: \"da1a9dd7-58b2-458e-8134-8251beec7d02\") " Dec 01 21:56:20 crc kubenswrapper[4857]: I1201 21:56:20.973525 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da1a9dd7-58b2-458e-8134-8251beec7d02-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "da1a9dd7-58b2-458e-8134-8251beec7d02" (UID: "da1a9dd7-58b2-458e-8134-8251beec7d02"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:20.994341 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da1a9dd7-58b2-458e-8134-8251beec7d02-logs" (OuterVolumeSpecName: "logs") pod "da1a9dd7-58b2-458e-8134-8251beec7d02" (UID: "da1a9dd7-58b2-458e-8134-8251beec7d02"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.014437 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-scripts" (OuterVolumeSpecName: "scripts") pod "da1a9dd7-58b2-458e-8134-8251beec7d02" (UID: "da1a9dd7-58b2-458e-8134-8251beec7d02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.016265 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da1a9dd7-58b2-458e-8134-8251beec7d02-kube-api-access-vpd85" (OuterVolumeSpecName: "kube-api-access-vpd85") pod "da1a9dd7-58b2-458e-8134-8251beec7d02" (UID: "da1a9dd7-58b2-458e-8134-8251beec7d02"). InnerVolumeSpecName "kube-api-access-vpd85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.035465 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "da1a9dd7-58b2-458e-8134-8251beec7d02" (UID: "da1a9dd7-58b2-458e-8134-8251beec7d02"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.075210 4857 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.075289 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.075309 4857 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da1a9dd7-58b2-458e-8134-8251beec7d02-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.075335 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpd85\" (UniqueName: \"kubernetes.io/projected/da1a9dd7-58b2-458e-8134-8251beec7d02-kube-api-access-vpd85\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.075378 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da1a9dd7-58b2-458e-8134-8251beec7d02-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.159921 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da1a9dd7-58b2-458e-8134-8251beec7d02" (UID: "da1a9dd7-58b2-458e-8134-8251beec7d02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.172508 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-config-data" (OuterVolumeSpecName: "config-data") pod "da1a9dd7-58b2-458e-8134-8251beec7d02" (UID: "da1a9dd7-58b2-458e-8134-8251beec7d02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.184292 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.184327 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da1a9dd7-58b2-458e-8134-8251beec7d02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.280000 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.298673 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.310806 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.390125 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-config-data\") pod \"07e01fec-a494-44c1-9671-b5f32f79e96a\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.390292 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tswh\" (UniqueName: \"kubernetes.io/projected/07e01fec-a494-44c1-9671-b5f32f79e96a-kube-api-access-6tswh\") pod \"07e01fec-a494-44c1-9671-b5f32f79e96a\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.390339 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07e01fec-a494-44c1-9671-b5f32f79e96a-httpd-run\") pod \"07e01fec-a494-44c1-9671-b5f32f79e96a\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.390381 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-config-data-custom\") pod \"69afe439-86e6-490f-9476-ebb89b540192\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.390408 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-combined-ca-bundle\") pod \"07e01fec-a494-44c1-9671-b5f32f79e96a\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.390433 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grnrb\" (UniqueName: \"kubernetes.io/projected/69afe439-86e6-490f-9476-ebb89b540192-kube-api-access-grnrb\") pod \"69afe439-86e6-490f-9476-ebb89b540192\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.390553 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-config-data\") pod \"69afe439-86e6-490f-9476-ebb89b540192\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.390638 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-combined-ca-bundle\") pod \"69afe439-86e6-490f-9476-ebb89b540192\" (UID: \"69afe439-86e6-490f-9476-ebb89b540192\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.390708 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-scripts\") pod \"07e01fec-a494-44c1-9671-b5f32f79e96a\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.390846 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-public-tls-certs\") pod \"07e01fec-a494-44c1-9671-b5f32f79e96a\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.390907 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"07e01fec-a494-44c1-9671-b5f32f79e96a\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.390929 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07e01fec-a494-44c1-9671-b5f32f79e96a-logs\") pod \"07e01fec-a494-44c1-9671-b5f32f79e96a\" (UID: \"07e01fec-a494-44c1-9671-b5f32f79e96a\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.392120 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07e01fec-a494-44c1-9671-b5f32f79e96a-logs" (OuterVolumeSpecName: "logs") pod "07e01fec-a494-44c1-9671-b5f32f79e96a" (UID: "07e01fec-a494-44c1-9671-b5f32f79e96a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.392157 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07e01fec-a494-44c1-9671-b5f32f79e96a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "07e01fec-a494-44c1-9671-b5f32f79e96a" (UID: "07e01fec-a494-44c1-9671-b5f32f79e96a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.403299 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69afe439-86e6-490f-9476-ebb89b540192-kube-api-access-grnrb" (OuterVolumeSpecName: "kube-api-access-grnrb") pod "69afe439-86e6-490f-9476-ebb89b540192" (UID: "69afe439-86e6-490f-9476-ebb89b540192"). InnerVolumeSpecName "kube-api-access-grnrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.404388 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-scripts" (OuterVolumeSpecName: "scripts") pod "07e01fec-a494-44c1-9671-b5f32f79e96a" (UID: "07e01fec-a494-44c1-9671-b5f32f79e96a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.404655 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "69afe439-86e6-490f-9476-ebb89b540192" (UID: "69afe439-86e6-490f-9476-ebb89b540192"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.404660 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "07e01fec-a494-44c1-9671-b5f32f79e96a" (UID: "07e01fec-a494-44c1-9671-b5f32f79e96a"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.431841 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07e01fec-a494-44c1-9671-b5f32f79e96a-kube-api-access-6tswh" (OuterVolumeSpecName: "kube-api-access-6tswh") pod "07e01fec-a494-44c1-9671-b5f32f79e96a" (UID: "07e01fec-a494-44c1-9671-b5f32f79e96a"). InnerVolumeSpecName "kube-api-access-6tswh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.439823 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07e01fec-a494-44c1-9671-b5f32f79e96a" (UID: "07e01fec-a494-44c1-9671-b5f32f79e96a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.461453 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69afe439-86e6-490f-9476-ebb89b540192" (UID: "69afe439-86e6-490f-9476-ebb89b540192"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.492842 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-config-data\") pod \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.493230 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-config-data-custom\") pod \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.493443 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xmzf\" (UniqueName: \"kubernetes.io/projected/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-kube-api-access-9xmzf\") pod \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.493575 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-combined-ca-bundle\") pod \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\" (UID: \"502899ed-eb9a-4eae-a3f3-1ef7369c08e7\") " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.494640 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.494973 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.495070 4857 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.496329 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07e01fec-a494-44c1-9671-b5f32f79e96a-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.497137 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tswh\" (UniqueName: \"kubernetes.io/projected/07e01fec-a494-44c1-9671-b5f32f79e96a-kube-api-access-6tswh\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.497409 4857 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07e01fec-a494-44c1-9671-b5f32f79e96a-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.497435 4857 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.497448 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.497464 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grnrb\" (UniqueName: \"kubernetes.io/projected/69afe439-86e6-490f-9476-ebb89b540192-kube-api-access-grnrb\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.539369 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-kube-api-access-9xmzf" (OuterVolumeSpecName: "kube-api-access-9xmzf") pod "502899ed-eb9a-4eae-a3f3-1ef7369c08e7" (UID: "502899ed-eb9a-4eae-a3f3-1ef7369c08e7"). InnerVolumeSpecName "kube-api-access-9xmzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.542168 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "502899ed-eb9a-4eae-a3f3-1ef7369c08e7" (UID: "502899ed-eb9a-4eae-a3f3-1ef7369c08e7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.550804 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dhc4w"] Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.559477 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d8e3-account-create-update-nrctx"] Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.621029 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" event={"ID":"69afe439-86e6-490f-9476-ebb89b540192","Type":"ContainerDied","Data":"bd8dc4423f6c5c2cd256329cf4bd0ca73544dccbfa97eac950a60d5c366c17bf"} Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.624105 4857 scope.go:117] "RemoveContainer" containerID="5739d985f701aa8a38cd890c6e351dcaf126790bc2552d67e58b2f5f879831e2" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.624056 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-75f4dbf68b-t5t5b" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.626318 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-config-data" (OuterVolumeSpecName: "config-data") pod "69afe439-86e6-490f-9476-ebb89b540192" (UID: "69afe439-86e6-490f-9476-ebb89b540192"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.626471 4857 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.626550 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xmzf\" (UniqueName: \"kubernetes.io/projected/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-kube-api-access-9xmzf\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.638109 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-config-data" (OuterVolumeSpecName: "config-data") pod "07e01fec-a494-44c1-9671-b5f32f79e96a" (UID: "07e01fec-a494-44c1-9671-b5f32f79e96a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.652156 4857 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.692371 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da1a9dd7-58b2-458e-8134-8251beec7d02","Type":"ContainerDied","Data":"182b4e022df02796ba488a9af226cbe150441e4f06afd68c36b9b7210036a3fa"} Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.692506 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.693406 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-75f4dbf68b-t5t5b"] Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.701764 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-75f4dbf68b-t5t5b"] Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.706244 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"07e01fec-a494-44c1-9671-b5f32f79e96a","Type":"ContainerDied","Data":"5cd726cebba3a46175fd824f89fad26e2ac79ed69a4a7746e2b1dd1c08f4628b"} Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.706339 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.710235 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77bf9494c9-4bdzk" event={"ID":"502899ed-eb9a-4eae-a3f3-1ef7369c08e7","Type":"ContainerDied","Data":"1820300bc5e769e194e42ec372a079e75bd1f5d88d4b389121ae6b15d8295769"} Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.710379 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77bf9494c9-4bdzk" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.713025 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913cd3df-6038-44e3-ad22-fc0fa6645058","Type":"ContainerStarted","Data":"014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c"} Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.734149 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.734187 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69afe439-86e6-490f-9476-ebb89b540192-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.734199 4857 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.777424 4857 scope.go:117] "RemoveContainer" containerID="b0e084fffedc0dcf7b68c91f7c8bcbbed5b83dcbcbb7b5f088756c5a4cbbd871" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.777745 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.795681 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.886868 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69afe439-86e6-490f-9476-ebb89b540192" path="/var/lib/kubelet/pods/69afe439-86e6-490f-9476-ebb89b540192/volumes" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.887495 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da1a9dd7-58b2-458e-8134-8251beec7d02" path="/var/lib/kubelet/pods/da1a9dd7-58b2-458e-8134-8251beec7d02/volumes" Dec 01 21:56:21 crc kubenswrapper[4857]: I1201 21:56:21.940451 4857 scope.go:117] "RemoveContainer" containerID="acb2dabf8bafc309791b5cce9c795324445611eb89691058d21620bed518c8c0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.007196 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "07e01fec-a494-44c1-9671-b5f32f79e96a" (UID: "07e01fec-a494-44c1-9671-b5f32f79e96a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.012853 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "502899ed-eb9a-4eae-a3f3-1ef7369c08e7" (UID: "502899ed-eb9a-4eae-a3f3-1ef7369c08e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.055750 4857 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07e01fec-a494-44c1-9671-b5f32f79e96a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.055783 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.055952 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-config-data" (OuterVolumeSpecName: "config-data") pod "502899ed-eb9a-4eae-a3f3-1ef7369c08e7" (UID: "502899ed-eb9a-4eae-a3f3-1ef7369c08e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:22 crc kubenswrapper[4857]: W1201 21:56:22.074496 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod574a17c8_2c54_41d2_8390_aff1f2220638.slice/crio-9ddb8cdef1745fe3437e78c60594a576703c73c54506e453b05357ae9be43c52 WatchSource:0}: Error finding container 9ddb8cdef1745fe3437e78c60594a576703c73c54506e453b05357ae9be43c52: Status 404 returned error can't find the container with id 9ddb8cdef1745fe3437e78c60594a576703c73c54506e453b05357ae9be43c52 Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.158188 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/502899ed-eb9a-4eae-a3f3-1ef7369c08e7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.202528 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 21:56:22 crc kubenswrapper[4857]: E1201 21:56:22.203543 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69afe439-86e6-490f-9476-ebb89b540192" containerName="heat-cfnapi" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.203562 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="69afe439-86e6-490f-9476-ebb89b540192" containerName="heat-cfnapi" Dec 01 21:56:22 crc kubenswrapper[4857]: E1201 21:56:22.203614 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="502899ed-eb9a-4eae-a3f3-1ef7369c08e7" containerName="heat-api" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.203621 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="502899ed-eb9a-4eae-a3f3-1ef7369c08e7" containerName="heat-api" Dec 01 21:56:22 crc kubenswrapper[4857]: E1201 21:56:22.203662 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da1a9dd7-58b2-458e-8134-8251beec7d02" containerName="cinder-api-log" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.203669 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="da1a9dd7-58b2-458e-8134-8251beec7d02" containerName="cinder-api-log" Dec 01 21:56:22 crc kubenswrapper[4857]: E1201 21:56:22.203685 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da1a9dd7-58b2-458e-8134-8251beec7d02" containerName="cinder-api" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.203692 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="da1a9dd7-58b2-458e-8134-8251beec7d02" containerName="cinder-api" Dec 01 21:56:22 crc kubenswrapper[4857]: E1201 21:56:22.203701 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07e01fec-a494-44c1-9671-b5f32f79e96a" containerName="glance-log" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.203708 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="07e01fec-a494-44c1-9671-b5f32f79e96a" containerName="glance-log" Dec 01 21:56:22 crc kubenswrapper[4857]: E1201 21:56:22.203731 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07e01fec-a494-44c1-9671-b5f32f79e96a" containerName="glance-httpd" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.203741 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="07e01fec-a494-44c1-9671-b5f32f79e96a" containerName="glance-httpd" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.203984 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="07e01fec-a494-44c1-9671-b5f32f79e96a" containerName="glance-log" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.205351 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="da1a9dd7-58b2-458e-8134-8251beec7d02" containerName="cinder-api" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.205367 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="da1a9dd7-58b2-458e-8134-8251beec7d02" containerName="cinder-api-log" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.205405 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="502899ed-eb9a-4eae-a3f3-1ef7369c08e7" containerName="heat-api" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.205413 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="69afe439-86e6-490f-9476-ebb89b540192" containerName="heat-cfnapi" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.205421 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="07e01fec-a494-44c1-9671-b5f32f79e96a" containerName="glance-httpd" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.207482 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.207546 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5894b69764-fpz5r" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.207562 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cjk48"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.207590 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-5c86f5f679-g6dgl" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.207606 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ktkxd"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.207617 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-cb18-account-create-update-q2fbk"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.207630 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7c37-account-create-update-rzsv8"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.208206 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.221958 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.223207 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.223338 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.260339 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-public-tls-certs\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.260702 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-config-data\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.260802 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-logs\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.260950 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-scripts\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.261028 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxm8c\" (UniqueName: \"kubernetes.io/projected/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-kube-api-access-vxm8c\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.261160 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.261271 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.261371 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.261481 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-config-data-custom\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.285975 4857 scope.go:117] "RemoveContainer" containerID="52086b2db0db0092223027f370cb55c8d6cbdd2a7c93d9296553b1a04101e5db" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.365442 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-config-data\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.365493 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-logs\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.365542 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-scripts\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.365564 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxm8c\" (UniqueName: \"kubernetes.io/projected/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-kube-api-access-vxm8c\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.365591 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.365612 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.365637 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.365666 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-config-data-custom\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.365719 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-public-tls-certs\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.378179 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-logs\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.384672 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.388639 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-66cd8df7d6-zlb8c"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.443111 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-config-data\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.451541 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxm8c\" (UniqueName: \"kubernetes.io/projected/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-kube-api-access-vxm8c\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.452367 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-public-tls-certs\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.502563 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-f87845c79-xkcmc"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.515007 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-scripts\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.552838 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.555016 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-config-data-custom\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.555107 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-77bf9494c9-4bdzk"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.561249 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/919f0ad8-05ea-44a1-a4ae-f81f4255e8da-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"919f0ad8-05ea-44a1-a4ae-f81f4255e8da\") " pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.601952 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-77bf9494c9-4bdzk"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.654907 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.657194 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.682583 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.760801 4857 scope.go:117] "RemoveContainer" containerID="4925b165c7077377e62bea2da2c5c31c88ce758e3a6407de1e148ddf0131b679" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.774434 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.784862 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.784912 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dhc4w" event={"ID":"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c","Type":"ContainerStarted","Data":"db0bb80baacfd9772df1c6bd3dcb59e080099536028d6a925ec7380dd6a5dc04"} Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.785010 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.802185 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.802295 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.803644 4857 scope.go:117] "RemoveContainer" containerID="b97342b2bd102c51865c503ee965467bd9493b0f742e791191d394191444f65b" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.832763 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ktkxd" event={"ID":"cc85532c-b753-4a79-ad36-621f558d03d0","Type":"ContainerStarted","Data":"cb3e7b7b76fa9c0745390b20ffce1e976f203941fbc2e75be6cca626b40c2c1f"} Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.863551 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6821156d-2d04-4608-a5c7-c70b1058fdc0","Type":"ContainerStarted","Data":"c657db47da331af1813f6ade9364da74d4954e848ff4cf4cb00cf13198f57d1d"} Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.889063 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cjk48" event={"ID":"d9e73cb1-7588-46b5-9014-efe779354e1d","Type":"ContainerStarted","Data":"aa46f0f86c334632691607fe94f4a2b9c537ada6d07a5bace42fb2d56885f463"} Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.892547 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-cb18-account-create-update-q2fbk" event={"ID":"a7174f1f-8255-4dfe-a230-5128ba63571e","Type":"ContainerStarted","Data":"75d778a85a65827d238e562e09a6f208de138d372aa0ac24159d5846802913dc"} Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.897082 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=4.978356822 podStartE2EDuration="29.897069648s" podCreationTimestamp="2025-12-01 21:55:53 +0000 UTC" firstStartedPulling="2025-12-01 21:55:55.281615171 +0000 UTC m=+1273.771677488" lastFinishedPulling="2025-12-01 21:56:20.200327997 +0000 UTC m=+1298.690390314" observedRunningTime="2025-12-01 21:56:22.894034044 +0000 UTC m=+1301.384096361" watchObservedRunningTime="2025-12-01 21:56:22.897069648 +0000 UTC m=+1301.387131965" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.918872 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" event={"ID":"574a17c8-2c54-41d2-8390-aff1f2220638","Type":"ContainerStarted","Data":"9ddb8cdef1745fe3437e78c60594a576703c73c54506e453b05357ae9be43c52"} Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.919908 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.920000 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.920177 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-logs\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.920225 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.920248 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wb5s\" (UniqueName: \"kubernetes.io/projected/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-kube-api-access-8wb5s\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.920293 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.920360 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.920447 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.941181 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-cb18-account-create-update-q2fbk" podStartSLOduration=6.941161749 podStartE2EDuration="6.941161749s" podCreationTimestamp="2025-12-01 21:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:22.919263072 +0000 UTC m=+1301.409325389" watchObservedRunningTime="2025-12-01 21:56:22.941161749 +0000 UTC m=+1301.431224066" Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.950510 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" event={"ID":"9638968d-42b2-46ab-b02a-72d044325fa9","Type":"ContainerStarted","Data":"dbb202fdc130888f71b8909accc0b79a3ea00018f73fdf434eee22ca8109ceb4"} Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.950576 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" event={"ID":"9638968d-42b2-46ab-b02a-72d044325fa9","Type":"ContainerStarted","Data":"f7527c479d9483d9e3e5ecbe2f2f58e7366891d4157d096d6d009153a696a891"} Dec 01 21:56:22 crc kubenswrapper[4857]: I1201 21:56:22.990343 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" podStartSLOduration=6.990328234 podStartE2EDuration="6.990328234s" podCreationTimestamp="2025-12-01 21:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:22.983376164 +0000 UTC m=+1301.473438481" watchObservedRunningTime="2025-12-01 21:56:22.990328234 +0000 UTC m=+1301.480390541" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.009397 4857 generic.go:334] "Generic (PLEG): container finished" podID="3ec1f836-90ab-4f44-b500-5a705a165925" containerID="0761be5c409a673752c97c6527eca4e0dabc3ce5b2e9374fa9db1ff742f67ccd" exitCode=0 Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.009475 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3ec1f836-90ab-4f44-b500-5a705a165925","Type":"ContainerDied","Data":"0761be5c409a673752c97c6527eca4e0dabc3ce5b2e9374fa9db1ff742f67ccd"} Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.035650 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.035705 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.035752 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.035838 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-logs\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.035872 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.035898 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wb5s\" (UniqueName: \"kubernetes.io/projected/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-kube-api-access-8wb5s\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.035939 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.036011 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.036536 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.037703 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-logs\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.038166 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.044787 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.045742 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.046142 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.046626 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.124303 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wb5s\" (UniqueName: \"kubernetes.io/projected/dc169c0b-67f2-48bf-8ac6-f58ca383ae4c-kube-api-access-8wb5s\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.236247 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c\") " pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.383809 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.460852 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-config-data\") pod \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.460917 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-combined-ca-bundle\") pod \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.462722 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58m6t\" (UniqueName: \"kubernetes.io/projected/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-kube-api-access-58m6t\") pod \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.462752 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-config-data-custom\") pod \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\" (UID: \"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24\") " Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.482633 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.497467 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" (UID: "9dff5cc8-16db-4b2b-8d0a-72eb2c068c24"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.498495 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-kube-api-access-58m6t" (OuterVolumeSpecName: "kube-api-access-58m6t") pod "9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" (UID: "9dff5cc8-16db-4b2b-8d0a-72eb2c068c24"). InnerVolumeSpecName "kube-api-access-58m6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.566681 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58m6t\" (UniqueName: \"kubernetes.io/projected/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-kube-api-access-58m6t\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.566734 4857 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.773656 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" (UID: "9dff5cc8-16db-4b2b-8d0a-72eb2c068c24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.874774 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.898286 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-config-data" (OuterVolumeSpecName: "config-data") pod "9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" (UID: "9dff5cc8-16db-4b2b-8d0a-72eb2c068c24"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.912115 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07e01fec-a494-44c1-9671-b5f32f79e96a" path="/var/lib/kubelet/pods/07e01fec-a494-44c1-9671-b5f32f79e96a/volumes" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.920822 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="502899ed-eb9a-4eae-a3f3-1ef7369c08e7" path="/var/lib/kubelet/pods/502899ed-eb9a-4eae-a3f3-1ef7369c08e7/volumes" Dec 01 21:56:23 crc kubenswrapper[4857]: W1201 21:56:23.977996 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod919f0ad8_05ea_44a1_a4ae_f81f4255e8da.slice/crio-f34758a24539e9d74c339d4bc6d04b458d26af629b2ca6b6fc08d0f83cbab81e WatchSource:0}: Error finding container f34758a24539e9d74c339d4bc6d04b458d26af629b2ca6b6fc08d0f83cbab81e: Status 404 returned error can't find the container with id f34758a24539e9d74c339d4bc6d04b458d26af629b2ca6b6fc08d0f83cbab81e Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.980401 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:23 crc kubenswrapper[4857]: I1201 21:56:23.988464 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.060547 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913cd3df-6038-44e3-ad22-fc0fa6645058","Type":"ContainerStarted","Data":"9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d"} Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.064087 4857 generic.go:334] "Generic (PLEG): container finished" podID="a7174f1f-8255-4dfe-a230-5128ba63571e" containerID="ad4dfe871a26d8c5d1ed8e6808025f3b4fc406625ed549bb498fb4d8cae5c9e9" exitCode=0 Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.064163 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-cb18-account-create-update-q2fbk" event={"ID":"a7174f1f-8255-4dfe-a230-5128ba63571e","Type":"ContainerDied","Data":"ad4dfe871a26d8c5d1ed8e6808025f3b4fc406625ed549bb498fb4d8cae5c9e9"} Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.069403 4857 generic.go:334] "Generic (PLEG): container finished" podID="16aaf417-9ea4-4f49-ad7b-77064d9fcb8c" containerID="f68ec44666a1cf9a01d3df410f654e5967c439d9ffc2cbdeb314b5120496b0e1" exitCode=0 Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.069516 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dhc4w" event={"ID":"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c","Type":"ContainerDied","Data":"f68ec44666a1cf9a01d3df410f654e5967c439d9ffc2cbdeb314b5120496b0e1"} Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.090267 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f87845c79-xkcmc" event={"ID":"e8524783-2889-48f9-b8f9-0d5fb3db15e5","Type":"ContainerDied","Data":"8efb290de50b48a4ca1ab7c6cca746b424c7a75463781c6111a8bb3709280477"} Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.090305 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8efb290de50b48a4ca1ab7c6cca746b424c7a75463781c6111a8bb3709280477" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.108294 4857 generic.go:334] "Generic (PLEG): container finished" podID="cc85532c-b753-4a79-ad36-621f558d03d0" containerID="fd5e469f0ee3459dff1e28e46ccefe6cbfd8604cbb8469fe8a30030c53c8c130" exitCode=0 Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.108388 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ktkxd" event={"ID":"cc85532c-b753-4a79-ad36-621f558d03d0","Type":"ContainerDied","Data":"fd5e469f0ee3459dff1e28e46ccefe6cbfd8604cbb8469fe8a30030c53c8c130"} Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.136541 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" event={"ID":"574a17c8-2c54-41d2-8390-aff1f2220638","Type":"ContainerStarted","Data":"78ff7551210d68af64ef40de04477bb2d7bd350b72c7497937d3602bc511379e"} Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.180904 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3ec1f836-90ab-4f44-b500-5a705a165925","Type":"ContainerDied","Data":"1808b55fd2ae872c67b19382b9fba0ef2118c61f9d83f14dfec98339ef26f83d"} Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.180947 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1808b55fd2ae872c67b19382b9fba0ef2118c61f9d83f14dfec98339ef26f83d" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.184154 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"919f0ad8-05ea-44a1-a4ae-f81f4255e8da","Type":"ContainerStarted","Data":"f34758a24539e9d74c339d4bc6d04b458d26af629b2ca6b6fc08d0f83cbab81e"} Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.206536 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cjk48" event={"ID":"d9e73cb1-7588-46b5-9014-efe779354e1d","Type":"ContainerStarted","Data":"3d5c157e719b635007d763a137b271d169fb1abf27dd50770eb6d1b582825668"} Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.245202 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-cjk48" podStartSLOduration=9.245184983 podStartE2EDuration="9.245184983s" podCreationTimestamp="2025-12-01 21:56:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:24.23569928 +0000 UTC m=+1302.725761597" watchObservedRunningTime="2025-12-01 21:56:24.245184983 +0000 UTC m=+1302.735247300" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.246774 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" event={"ID":"9dff5cc8-16db-4b2b-8d0a-72eb2c068c24","Type":"ContainerDied","Data":"053443615a3a777f1e3150c36f425d5ee7177b6b4ef7fd4e32213fed40524944"} Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.246816 4857 scope.go:117] "RemoveContainer" containerID="e63cb4ca5c7b1ea70185621cc0ce903fcf6418219b29fd3166b264329ce01a47" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.246895 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66cd8df7d6-zlb8c" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.253681 4857 generic.go:334] "Generic (PLEG): container finished" podID="9638968d-42b2-46ab-b02a-72d044325fa9" containerID="dbb202fdc130888f71b8909accc0b79a3ea00018f73fdf434eee22ca8109ceb4" exitCode=0 Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.253850 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" event={"ID":"9638968d-42b2-46ab-b02a-72d044325fa9","Type":"ContainerDied","Data":"dbb202fdc130888f71b8909accc0b79a3ea00018f73fdf434eee22ca8109ceb4"} Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.381450 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.438519 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.444188 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.446206 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-66cd8df7d6-zlb8c"] Dec 01 21:56:24 crc kubenswrapper[4857]: W1201 21:56:24.464035 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc169c0b_67f2_48bf_8ac6_f58ca383ae4c.slice/crio-87eb2c5702adfc6b7cad5b03f5100bfe7eaa7d6f2a11be27e201a85bb9b8c681 WatchSource:0}: Error finding container 87eb2c5702adfc6b7cad5b03f5100bfe7eaa7d6f2a11be27e201a85bb9b8c681: Status 404 returned error can't find the container with id 87eb2c5702adfc6b7cad5b03f5100bfe7eaa7d6f2a11be27e201a85bb9b8c681 Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.473131 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-66cd8df7d6-zlb8c"] Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.608887 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ec1f836-90ab-4f44-b500-5a705a165925-httpd-run\") pod \"3ec1f836-90ab-4f44-b500-5a705a165925\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.608954 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-scripts\") pod \"3ec1f836-90ab-4f44-b500-5a705a165925\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.608979 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-combined-ca-bundle\") pod \"3ec1f836-90ab-4f44-b500-5a705a165925\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.609008 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwwv7\" (UniqueName: \"kubernetes.io/projected/e8524783-2889-48f9-b8f9-0d5fb3db15e5-kube-api-access-mwwv7\") pod \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.609089 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ec1f836-90ab-4f44-b500-5a705a165925-logs\") pod \"3ec1f836-90ab-4f44-b500-5a705a165925\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.609122 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-internal-tls-certs\") pod \"3ec1f836-90ab-4f44-b500-5a705a165925\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.609152 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-config-data-custom\") pod \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.609198 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"3ec1f836-90ab-4f44-b500-5a705a165925\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.609234 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-config-data\") pod \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.609257 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-config-data\") pod \"3ec1f836-90ab-4f44-b500-5a705a165925\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.609401 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-combined-ca-bundle\") pod \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.609422 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9fxr\" (UniqueName: \"kubernetes.io/projected/3ec1f836-90ab-4f44-b500-5a705a165925-kube-api-access-m9fxr\") pod \"3ec1f836-90ab-4f44-b500-5a705a165925\" (UID: \"3ec1f836-90ab-4f44-b500-5a705a165925\") " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.613340 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ec1f836-90ab-4f44-b500-5a705a165925-logs" (OuterVolumeSpecName: "logs") pod "3ec1f836-90ab-4f44-b500-5a705a165925" (UID: "3ec1f836-90ab-4f44-b500-5a705a165925"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.616625 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ec1f836-90ab-4f44-b500-5a705a165925-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3ec1f836-90ab-4f44-b500-5a705a165925" (UID: "3ec1f836-90ab-4f44-b500-5a705a165925"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.621750 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-scripts" (OuterVolumeSpecName: "scripts") pod "3ec1f836-90ab-4f44-b500-5a705a165925" (UID: "3ec1f836-90ab-4f44-b500-5a705a165925"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.624225 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec1f836-90ab-4f44-b500-5a705a165925-kube-api-access-m9fxr" (OuterVolumeSpecName: "kube-api-access-m9fxr") pod "3ec1f836-90ab-4f44-b500-5a705a165925" (UID: "3ec1f836-90ab-4f44-b500-5a705a165925"). InnerVolumeSpecName "kube-api-access-m9fxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.626645 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "3ec1f836-90ab-4f44-b500-5a705a165925" (UID: "3ec1f836-90ab-4f44-b500-5a705a165925"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.631688 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8524783-2889-48f9-b8f9-0d5fb3db15e5-kube-api-access-mwwv7" (OuterVolumeSpecName: "kube-api-access-mwwv7") pod "e8524783-2889-48f9-b8f9-0d5fb3db15e5" (UID: "e8524783-2889-48f9-b8f9-0d5fb3db15e5"). InnerVolumeSpecName "kube-api-access-mwwv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.662278 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e8524783-2889-48f9-b8f9-0d5fb3db15e5" (UID: "e8524783-2889-48f9-b8f9-0d5fb3db15e5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.686962 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ec1f836-90ab-4f44-b500-5a705a165925" (UID: "3ec1f836-90ab-4f44-b500-5a705a165925"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.714902 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8524783-2889-48f9-b8f9-0d5fb3db15e5" (UID: "e8524783-2889-48f9-b8f9-0d5fb3db15e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.717949 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-combined-ca-bundle\") pod \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\" (UID: \"e8524783-2889-48f9-b8f9-0d5fb3db15e5\") " Dec 01 21:56:24 crc kubenswrapper[4857]: W1201 21:56:24.718090 4857 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/e8524783-2889-48f9-b8f9-0d5fb3db15e5/volumes/kubernetes.io~secret/combined-ca-bundle Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.718811 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8524783-2889-48f9-b8f9-0d5fb3db15e5" (UID: "e8524783-2889-48f9-b8f9-0d5fb3db15e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.721693 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-config-data" (OuterVolumeSpecName: "config-data") pod "e8524783-2889-48f9-b8f9-0d5fb3db15e5" (UID: "e8524783-2889-48f9-b8f9-0d5fb3db15e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.723171 4857 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ec1f836-90ab-4f44-b500-5a705a165925-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.723200 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.723212 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.723225 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwwv7\" (UniqueName: \"kubernetes.io/projected/e8524783-2889-48f9-b8f9-0d5fb3db15e5-kube-api-access-mwwv7\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.723239 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ec1f836-90ab-4f44-b500-5a705a165925-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.723249 4857 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.723284 4857 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.723295 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.723306 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8524783-2889-48f9-b8f9-0d5fb3db15e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.723317 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9fxr\" (UniqueName: \"kubernetes.io/projected/3ec1f836-90ab-4f44-b500-5a705a165925-kube-api-access-m9fxr\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.734133 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3ec1f836-90ab-4f44-b500-5a705a165925" (UID: "3ec1f836-90ab-4f44-b500-5a705a165925"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.761741 4857 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.780181 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-config-data" (OuterVolumeSpecName: "config-data") pod "3ec1f836-90ab-4f44-b500-5a705a165925" (UID: "3ec1f836-90ab-4f44-b500-5a705a165925"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.825611 4857 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.825655 4857 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:24 crc kubenswrapper[4857]: I1201 21:56:24.825668 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ec1f836-90ab-4f44-b500-5a705a165925-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.298798 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7c64685c4b-rgnfp" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.397386 4857 generic.go:334] "Generic (PLEG): container finished" podID="d9e73cb1-7588-46b5-9014-efe779354e1d" containerID="3d5c157e719b635007d763a137b271d169fb1abf27dd50770eb6d1b582825668" exitCode=0 Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.398038 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cjk48" event={"ID":"d9e73cb1-7588-46b5-9014-efe779354e1d","Type":"ContainerDied","Data":"3d5c157e719b635007d763a137b271d169fb1abf27dd50770eb6d1b582825668"} Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.407487 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c","Type":"ContainerStarted","Data":"87eb2c5702adfc6b7cad5b03f5100bfe7eaa7d6f2a11be27e201a85bb9b8c681"} Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.421022 4857 generic.go:334] "Generic (PLEG): container finished" podID="574a17c8-2c54-41d2-8390-aff1f2220638" containerID="78ff7551210d68af64ef40de04477bb2d7bd350b72c7497937d3602bc511379e" exitCode=0 Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.429532 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.420028 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-56454df68d-6lxzh"] Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.431184 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" event={"ID":"574a17c8-2c54-41d2-8390-aff1f2220638","Type":"ContainerDied","Data":"78ff7551210d68af64ef40de04477bb2d7bd350b72c7497937d3602bc511379e"} Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.431230 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"919f0ad8-05ea-44a1-a4ae-f81f4255e8da","Type":"ContainerStarted","Data":"f85916c06445fc3aeb41e88783bad26282b552ed068ffb41c2b73c69b2b7c4c0"} Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.431635 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-56454df68d-6lxzh" podUID="2842d110-9e75-4911-a86e-29327b79b8f8" containerName="heat-engine" containerID="cri-o://621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b" gracePeriod=60 Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.433256 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f87845c79-xkcmc" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.542714 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.566815 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.590167 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-f87845c79-xkcmc"] Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.623711 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-f87845c79-xkcmc"] Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650214 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:56:25 crc kubenswrapper[4857]: E1201 21:56:25.650662 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ec1f836-90ab-4f44-b500-5a705a165925" containerName="glance-log" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650675 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ec1f836-90ab-4f44-b500-5a705a165925" containerName="glance-log" Dec 01 21:56:25 crc kubenswrapper[4857]: E1201 21:56:25.650696 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8524783-2889-48f9-b8f9-0d5fb3db15e5" containerName="heat-api" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650701 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8524783-2889-48f9-b8f9-0d5fb3db15e5" containerName="heat-api" Dec 01 21:56:25 crc kubenswrapper[4857]: E1201 21:56:25.650715 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" containerName="heat-cfnapi" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650722 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" containerName="heat-cfnapi" Dec 01 21:56:25 crc kubenswrapper[4857]: E1201 21:56:25.650734 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ec1f836-90ab-4f44-b500-5a705a165925" containerName="glance-httpd" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650740 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ec1f836-90ab-4f44-b500-5a705a165925" containerName="glance-httpd" Dec 01 21:56:25 crc kubenswrapper[4857]: E1201 21:56:25.650746 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8524783-2889-48f9-b8f9-0d5fb3db15e5" containerName="heat-api" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650752 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8524783-2889-48f9-b8f9-0d5fb3db15e5" containerName="heat-api" Dec 01 21:56:25 crc kubenswrapper[4857]: E1201 21:56:25.650763 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" containerName="heat-cfnapi" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650769 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" containerName="heat-cfnapi" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650937 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" containerName="heat-cfnapi" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650949 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" containerName="heat-cfnapi" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650962 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ec1f836-90ab-4f44-b500-5a705a165925" containerName="glance-log" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650972 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ec1f836-90ab-4f44-b500-5a705a165925" containerName="glance-httpd" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650981 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8524783-2889-48f9-b8f9-0d5fb3db15e5" containerName="heat-api" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.650991 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8524783-2889-48f9-b8f9-0d5fb3db15e5" containerName="heat-api" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.651892 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.658265 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.658488 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.661355 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.864752 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.865322 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-logs\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.865368 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.865444 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.865473 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.865529 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.865577 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.865603 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbdgg\" (UniqueName: \"kubernetes.io/projected/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-kube-api-access-xbdgg\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.869026 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ec1f836-90ab-4f44-b500-5a705a165925" path="/var/lib/kubelet/pods/3ec1f836-90ab-4f44-b500-5a705a165925/volumes" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.870425 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dff5cc8-16db-4b2b-8d0a-72eb2c068c24" path="/var/lib/kubelet/pods/9dff5cc8-16db-4b2b-8d0a-72eb2c068c24/volumes" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.880790 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8524783-2889-48f9-b8f9-0d5fb3db15e5" path="/var/lib/kubelet/pods/e8524783-2889-48f9-b8f9-0d5fb3db15e5/volumes" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.984987 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.985081 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.985163 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.985225 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.985249 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbdgg\" (UniqueName: \"kubernetes.io/projected/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-kube-api-access-xbdgg\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.985295 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.985332 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-logs\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.985375 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.989018 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.989138 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:25 crc kubenswrapper[4857]: I1201 21:56:25.990434 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-logs\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.007591 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.031557 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.033736 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.039266 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbdgg\" (UniqueName: \"kubernetes.io/projected/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-kube-api-access-xbdgg\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.040330 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3b7eba-7e44-4f50-a2da-626cfe8e3921-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.043177 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"4f3b7eba-7e44-4f50-a2da-626cfe8e3921\") " pod="openstack/glance-default-internal-api-0" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.149377 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ktkxd" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.281455 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.289781 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtww2\" (UniqueName: \"kubernetes.io/projected/cc85532c-b753-4a79-ad36-621f558d03d0-kube-api-access-vtww2\") pod \"cc85532c-b753-4a79-ad36-621f558d03d0\" (UID: \"cc85532c-b753-4a79-ad36-621f558d03d0\") " Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.289939 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc85532c-b753-4a79-ad36-621f558d03d0-operator-scripts\") pod \"cc85532c-b753-4a79-ad36-621f558d03d0\" (UID: \"cc85532c-b753-4a79-ad36-621f558d03d0\") " Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.291737 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc85532c-b753-4a79-ad36-621f558d03d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cc85532c-b753-4a79-ad36-621f558d03d0" (UID: "cc85532c-b753-4a79-ad36-621f558d03d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.308113 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc85532c-b753-4a79-ad36-621f558d03d0-kube-api-access-vtww2" (OuterVolumeSpecName: "kube-api-access-vtww2") pod "cc85532c-b753-4a79-ad36-621f558d03d0" (UID: "cc85532c-b753-4a79-ad36-621f558d03d0"). InnerVolumeSpecName "kube-api-access-vtww2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.411171 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtww2\" (UniqueName: \"kubernetes.io/projected/cc85532c-b753-4a79-ad36-621f558d03d0-kube-api-access-vtww2\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.411201 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc85532c-b753-4a79-ad36-621f558d03d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.482216 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.484843 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dhc4w" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.516669 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmqs8\" (UniqueName: \"kubernetes.io/projected/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c-kube-api-access-jmqs8\") pod \"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c\" (UID: \"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c\") " Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.516782 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwswz\" (UniqueName: \"kubernetes.io/projected/574a17c8-2c54-41d2-8390-aff1f2220638-kube-api-access-dwswz\") pod \"574a17c8-2c54-41d2-8390-aff1f2220638\" (UID: \"574a17c8-2c54-41d2-8390-aff1f2220638\") " Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.516823 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c-operator-scripts\") pod \"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c\" (UID: \"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c\") " Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.516840 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/574a17c8-2c54-41d2-8390-aff1f2220638-operator-scripts\") pod \"574a17c8-2c54-41d2-8390-aff1f2220638\" (UID: \"574a17c8-2c54-41d2-8390-aff1f2220638\") " Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.517507 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/574a17c8-2c54-41d2-8390-aff1f2220638-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "574a17c8-2c54-41d2-8390-aff1f2220638" (UID: "574a17c8-2c54-41d2-8390-aff1f2220638"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.517912 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16aaf417-9ea4-4f49-ad7b-77064d9fcb8c" (UID: "16aaf417-9ea4-4f49-ad7b-77064d9fcb8c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.537841 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/574a17c8-2c54-41d2-8390-aff1f2220638-kube-api-access-dwswz" (OuterVolumeSpecName: "kube-api-access-dwswz") pod "574a17c8-2c54-41d2-8390-aff1f2220638" (UID: "574a17c8-2c54-41d2-8390-aff1f2220638"). InnerVolumeSpecName "kube-api-access-dwswz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.552388 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c-kube-api-access-jmqs8" (OuterVolumeSpecName: "kube-api-access-jmqs8") pod "16aaf417-9ea4-4f49-ad7b-77064d9fcb8c" (UID: "16aaf417-9ea4-4f49-ad7b-77064d9fcb8c"). InnerVolumeSpecName "kube-api-access-jmqs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.560877 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" event={"ID":"574a17c8-2c54-41d2-8390-aff1f2220638","Type":"ContainerDied","Data":"9ddb8cdef1745fe3437e78c60594a576703c73c54506e453b05357ae9be43c52"} Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.560921 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ddb8cdef1745fe3437e78c60594a576703c73c54506e453b05357ae9be43c52" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.560994 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7c37-account-create-update-rzsv8" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.569075 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-cb18-account-create-update-q2fbk" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.570148 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ktkxd" event={"ID":"cc85532c-b753-4a79-ad36-621f558d03d0","Type":"ContainerDied","Data":"cb3e7b7b76fa9c0745390b20ffce1e976f203941fbc2e75be6cca626b40c2c1f"} Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.570178 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb3e7b7b76fa9c0745390b20ffce1e976f203941fbc2e75be6cca626b40c2c1f" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.570222 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ktkxd" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.618328 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7174f1f-8255-4dfe-a230-5128ba63571e-operator-scripts\") pod \"a7174f1f-8255-4dfe-a230-5128ba63571e\" (UID: \"a7174f1f-8255-4dfe-a230-5128ba63571e\") " Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.618389 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttr4d\" (UniqueName: \"kubernetes.io/projected/a7174f1f-8255-4dfe-a230-5128ba63571e-kube-api-access-ttr4d\") pod \"a7174f1f-8255-4dfe-a230-5128ba63571e\" (UID: \"a7174f1f-8255-4dfe-a230-5128ba63571e\") " Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.618628 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwswz\" (UniqueName: \"kubernetes.io/projected/574a17c8-2c54-41d2-8390-aff1f2220638-kube-api-access-dwswz\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.618639 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.618648 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/574a17c8-2c54-41d2-8390-aff1f2220638-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.618656 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmqs8\" (UniqueName: \"kubernetes.io/projected/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c-kube-api-access-jmqs8\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.623443 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7174f1f-8255-4dfe-a230-5128ba63571e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a7174f1f-8255-4dfe-a230-5128ba63571e" (UID: "a7174f1f-8255-4dfe-a230-5128ba63571e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.623571 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7174f1f-8255-4dfe-a230-5128ba63571e-kube-api-access-ttr4d" (OuterVolumeSpecName: "kube-api-access-ttr4d") pod "a7174f1f-8255-4dfe-a230-5128ba63571e" (UID: "a7174f1f-8255-4dfe-a230-5128ba63571e"). InnerVolumeSpecName "kube-api-access-ttr4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.624644 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.625836 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913cd3df-6038-44e3-ad22-fc0fa6645058","Type":"ContainerStarted","Data":"6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440"} Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.625980 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="ceilometer-central-agent" containerID="cri-o://d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1" gracePeriod=30 Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.626198 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.626237 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="proxy-httpd" containerID="cri-o://6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440" gracePeriod=30 Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.626281 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="sg-core" containerID="cri-o://9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d" gracePeriod=30 Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.626312 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="ceilometer-notification-agent" containerID="cri-o://014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c" gracePeriod=30 Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.635524 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dhc4w" event={"ID":"16aaf417-9ea4-4f49-ad7b-77064d9fcb8c","Type":"ContainerDied","Data":"db0bb80baacfd9772df1c6bd3dcb59e080099536028d6a925ec7380dd6a5dc04"} Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.635562 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db0bb80baacfd9772df1c6bd3dcb59e080099536028d6a925ec7380dd6a5dc04" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.635623 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dhc4w" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.662134 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c","Type":"ContainerStarted","Data":"66c3702c5402534683525f2f9e47021ec0b0f84386fe28774ade6d3c2d67cc64"} Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.707184 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.514280397 podStartE2EDuration="18.70716682s" podCreationTimestamp="2025-12-01 21:56:08 +0000 UTC" firstStartedPulling="2025-12-01 21:56:09.356208071 +0000 UTC m=+1287.846270388" lastFinishedPulling="2025-12-01 21:56:25.549094494 +0000 UTC m=+1304.039156811" observedRunningTime="2025-12-01 21:56:26.706381691 +0000 UTC m=+1305.196444008" watchObservedRunningTime="2025-12-01 21:56:26.70716682 +0000 UTC m=+1305.197229137" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.719014 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9638968d-42b2-46ab-b02a-72d044325fa9-operator-scripts\") pod \"9638968d-42b2-46ab-b02a-72d044325fa9\" (UID: \"9638968d-42b2-46ab-b02a-72d044325fa9\") " Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.719149 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bw6j\" (UniqueName: \"kubernetes.io/projected/9638968d-42b2-46ab-b02a-72d044325fa9-kube-api-access-8bw6j\") pod \"9638968d-42b2-46ab-b02a-72d044325fa9\" (UID: \"9638968d-42b2-46ab-b02a-72d044325fa9\") " Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.719440 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7174f1f-8255-4dfe-a230-5128ba63571e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.719458 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttr4d\" (UniqueName: \"kubernetes.io/projected/a7174f1f-8255-4dfe-a230-5128ba63571e-kube-api-access-ttr4d\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.722364 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9638968d-42b2-46ab-b02a-72d044325fa9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9638968d-42b2-46ab-b02a-72d044325fa9" (UID: "9638968d-42b2-46ab-b02a-72d044325fa9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.731427 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9638968d-42b2-46ab-b02a-72d044325fa9-kube-api-access-8bw6j" (OuterVolumeSpecName: "kube-api-access-8bw6j") pod "9638968d-42b2-46ab-b02a-72d044325fa9" (UID: "9638968d-42b2-46ab-b02a-72d044325fa9"). InnerVolumeSpecName "kube-api-access-8bw6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.823610 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bw6j\" (UniqueName: \"kubernetes.io/projected/9638968d-42b2-46ab-b02a-72d044325fa9-kube-api-access-8bw6j\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:26 crc kubenswrapper[4857]: I1201 21:56:26.823661 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9638968d-42b2-46ab-b02a-72d044325fa9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.138764 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cjk48" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.335117 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxgcw\" (UniqueName: \"kubernetes.io/projected/d9e73cb1-7588-46b5-9014-efe779354e1d-kube-api-access-nxgcw\") pod \"d9e73cb1-7588-46b5-9014-efe779354e1d\" (UID: \"d9e73cb1-7588-46b5-9014-efe779354e1d\") " Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.335281 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9e73cb1-7588-46b5-9014-efe779354e1d-operator-scripts\") pod \"d9e73cb1-7588-46b5-9014-efe779354e1d\" (UID: \"d9e73cb1-7588-46b5-9014-efe779354e1d\") " Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.336868 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9e73cb1-7588-46b5-9014-efe779354e1d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d9e73cb1-7588-46b5-9014-efe779354e1d" (UID: "d9e73cb1-7588-46b5-9014-efe779354e1d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.347898 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9e73cb1-7588-46b5-9014-efe779354e1d-kube-api-access-nxgcw" (OuterVolumeSpecName: "kube-api-access-nxgcw") pod "d9e73cb1-7588-46b5-9014-efe779354e1d" (UID: "d9e73cb1-7588-46b5-9014-efe779354e1d"). InnerVolumeSpecName "kube-api-access-nxgcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.418005 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 21:56:27 crc kubenswrapper[4857]: W1201 21:56:27.420110 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f3b7eba_7e44_4f50_a2da_626cfe8e3921.slice/crio-9ea65fc718d7b8d481d087cb36ae542e362186284e43f7f7cb1652cd73fc1d69 WatchSource:0}: Error finding container 9ea65fc718d7b8d481d087cb36ae542e362186284e43f7f7cb1652cd73fc1d69: Status 404 returned error can't find the container with id 9ea65fc718d7b8d481d087cb36ae542e362186284e43f7f7cb1652cd73fc1d69 Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.437615 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxgcw\" (UniqueName: \"kubernetes.io/projected/d9e73cb1-7588-46b5-9014-efe779354e1d-kube-api-access-nxgcw\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.437647 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9e73cb1-7588-46b5-9014-efe779354e1d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.678532 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cjk48" event={"ID":"d9e73cb1-7588-46b5-9014-efe779354e1d","Type":"ContainerDied","Data":"aa46f0f86c334632691607fe94f4a2b9c537ada6d07a5bace42fb2d56885f463"} Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.678576 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa46f0f86c334632691607fe94f4a2b9c537ada6d07a5bace42fb2d56885f463" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.678615 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cjk48" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.683926 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-cb18-account-create-update-q2fbk" event={"ID":"a7174f1f-8255-4dfe-a230-5128ba63571e","Type":"ContainerDied","Data":"75d778a85a65827d238e562e09a6f208de138d372aa0ac24159d5846802913dc"} Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.683965 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75d778a85a65827d238e562e09a6f208de138d372aa0ac24159d5846802913dc" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.684032 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-cb18-account-create-update-q2fbk" Dec 01 21:56:27 crc kubenswrapper[4857]: E1201 21:56:27.695762 4857 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.710996 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc169c0b-67f2-48bf-8ac6-f58ca383ae4c","Type":"ContainerStarted","Data":"d05d242c0e4916969f593f2a9b832443a7fd7b51dfd1975c66a839b5fa2398b9"} Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.721684 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" event={"ID":"9638968d-42b2-46ab-b02a-72d044325fa9","Type":"ContainerDied","Data":"f7527c479d9483d9e3e5ecbe2f2f58e7366891d4157d096d6d009153a696a891"} Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.721728 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7527c479d9483d9e3e5ecbe2f2f58e7366891d4157d096d6d009153a696a891" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.721786 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d8e3-account-create-update-nrctx" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.723812 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4f3b7eba-7e44-4f50-a2da-626cfe8e3921","Type":"ContainerStarted","Data":"9ea65fc718d7b8d481d087cb36ae542e362186284e43f7f7cb1652cd73fc1d69"} Dec 01 21:56:27 crc kubenswrapper[4857]: E1201 21:56:27.735137 4857 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.739954 4857 generic.go:334] "Generic (PLEG): container finished" podID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerID="9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d" exitCode=2 Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.740087 4857 generic.go:334] "Generic (PLEG): container finished" podID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerID="014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c" exitCode=0 Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.740148 4857 generic.go:334] "Generic (PLEG): container finished" podID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerID="d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1" exitCode=0 Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.740001 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913cd3df-6038-44e3-ad22-fc0fa6645058","Type":"ContainerDied","Data":"9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d"} Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.740319 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913cd3df-6038-44e3-ad22-fc0fa6645058","Type":"ContainerDied","Data":"014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c"} Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.740390 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913cd3df-6038-44e3-ad22-fc0fa6645058","Type":"ContainerDied","Data":"d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1"} Dec 01 21:56:27 crc kubenswrapper[4857]: E1201 21:56:27.740384 4857 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 01 21:56:27 crc kubenswrapper[4857]: E1201 21:56:27.743148 4857 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-56454df68d-6lxzh" podUID="2842d110-9e75-4911-a86e-29327b79b8f8" containerName="heat-engine" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.752673 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.752655667 podStartE2EDuration="5.752655667s" podCreationTimestamp="2025-12-01 21:56:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:27.750493414 +0000 UTC m=+1306.240555731" watchObservedRunningTime="2025-12-01 21:56:27.752655667 +0000 UTC m=+1306.242717984" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.758313 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"919f0ad8-05ea-44a1-a4ae-f81f4255e8da","Type":"ContainerStarted","Data":"ee5c4e12ee1522044f2ddb11886deb9114dcbd4fbb0c8075651135eb04d1d634"} Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.758478 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 21:56:27 crc kubenswrapper[4857]: I1201 21:56:27.800941 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.80092273 podStartE2EDuration="6.80092273s" podCreationTimestamp="2025-12-01 21:56:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:27.782237762 +0000 UTC m=+1306.272300079" watchObservedRunningTime="2025-12-01 21:56:27.80092273 +0000 UTC m=+1306.290985047" Dec 01 21:56:28 crc kubenswrapper[4857]: I1201 21:56:28.771344 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4f3b7eba-7e44-4f50-a2da-626cfe8e3921","Type":"ContainerStarted","Data":"5232b6975cc159791b43f237ee40f1c7f0a36795d266dd1201802eb79a010aad"} Dec 01 21:56:28 crc kubenswrapper[4857]: I1201 21:56:28.772146 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4f3b7eba-7e44-4f50-a2da-626cfe8e3921","Type":"ContainerStarted","Data":"643b52c46c68281cccfb96fb0c1440ca210f031aebf25f4a65ef2dfb14f2321b"} Dec 01 21:56:28 crc kubenswrapper[4857]: I1201 21:56:28.803847 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.803830192 podStartE2EDuration="3.803830192s" podCreationTimestamp="2025-12-01 21:56:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:56:28.788976918 +0000 UTC m=+1307.279039235" watchObservedRunningTime="2025-12-01 21:56:28.803830192 +0000 UTC m=+1307.293892519" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.776858 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-r7zf9"] Dec 01 21:56:31 crc kubenswrapper[4857]: E1201 21:56:31.777616 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574a17c8-2c54-41d2-8390-aff1f2220638" containerName="mariadb-account-create-update" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.777630 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="574a17c8-2c54-41d2-8390-aff1f2220638" containerName="mariadb-account-create-update" Dec 01 21:56:31 crc kubenswrapper[4857]: E1201 21:56:31.777642 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7174f1f-8255-4dfe-a230-5128ba63571e" containerName="mariadb-account-create-update" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.777648 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7174f1f-8255-4dfe-a230-5128ba63571e" containerName="mariadb-account-create-update" Dec 01 21:56:31 crc kubenswrapper[4857]: E1201 21:56:31.777666 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9638968d-42b2-46ab-b02a-72d044325fa9" containerName="mariadb-account-create-update" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.777672 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9638968d-42b2-46ab-b02a-72d044325fa9" containerName="mariadb-account-create-update" Dec 01 21:56:31 crc kubenswrapper[4857]: E1201 21:56:31.777684 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9e73cb1-7588-46b5-9014-efe779354e1d" containerName="mariadb-database-create" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.777689 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9e73cb1-7588-46b5-9014-efe779354e1d" containerName="mariadb-database-create" Dec 01 21:56:31 crc kubenswrapper[4857]: E1201 21:56:31.777699 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc85532c-b753-4a79-ad36-621f558d03d0" containerName="mariadb-database-create" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.777705 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc85532c-b753-4a79-ad36-621f558d03d0" containerName="mariadb-database-create" Dec 01 21:56:31 crc kubenswrapper[4857]: E1201 21:56:31.777722 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16aaf417-9ea4-4f49-ad7b-77064d9fcb8c" containerName="mariadb-database-create" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.777728 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="16aaf417-9ea4-4f49-ad7b-77064d9fcb8c" containerName="mariadb-database-create" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.777894 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9e73cb1-7588-46b5-9014-efe779354e1d" containerName="mariadb-database-create" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.777903 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="574a17c8-2c54-41d2-8390-aff1f2220638" containerName="mariadb-account-create-update" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.777915 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc85532c-b753-4a79-ad36-621f558d03d0" containerName="mariadb-database-create" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.777924 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="16aaf417-9ea4-4f49-ad7b-77064d9fcb8c" containerName="mariadb-database-create" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.777938 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9638968d-42b2-46ab-b02a-72d044325fa9" containerName="mariadb-account-create-update" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.777956 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7174f1f-8255-4dfe-a230-5128ba63571e" containerName="mariadb-account-create-update" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.778570 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.782883 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.783098 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.783301 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fc42k" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.816845 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-r7zf9"] Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.949332 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-config-data\") pod \"nova-cell0-conductor-db-sync-r7zf9\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.949439 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn6ht\" (UniqueName: \"kubernetes.io/projected/0906fcdc-a402-46cf-bf71-fcd7543a9034-kube-api-access-gn6ht\") pod \"nova-cell0-conductor-db-sync-r7zf9\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.949536 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-r7zf9\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:31 crc kubenswrapper[4857]: I1201 21:56:31.949591 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-scripts\") pod \"nova-cell0-conductor-db-sync-r7zf9\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.052168 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-r7zf9\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.052752 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-scripts\") pod \"nova-cell0-conductor-db-sync-r7zf9\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.052844 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-config-data\") pod \"nova-cell0-conductor-db-sync-r7zf9\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.052889 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn6ht\" (UniqueName: \"kubernetes.io/projected/0906fcdc-a402-46cf-bf71-fcd7543a9034-kube-api-access-gn6ht\") pod \"nova-cell0-conductor-db-sync-r7zf9\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.061778 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-config-data\") pod \"nova-cell0-conductor-db-sync-r7zf9\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.068554 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-scripts\") pod \"nova-cell0-conductor-db-sync-r7zf9\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.070016 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-r7zf9\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.073255 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn6ht\" (UniqueName: \"kubernetes.io/projected/0906fcdc-a402-46cf-bf71-fcd7543a9034-kube-api-access-gn6ht\") pod \"nova-cell0-conductor-db-sync-r7zf9\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.095619 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.398457 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.562857 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-config-data-custom\") pod \"2842d110-9e75-4911-a86e-29327b79b8f8\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.564692 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-combined-ca-bundle\") pod \"2842d110-9e75-4911-a86e-29327b79b8f8\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.564843 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j67zx\" (UniqueName: \"kubernetes.io/projected/2842d110-9e75-4911-a86e-29327b79b8f8-kube-api-access-j67zx\") pod \"2842d110-9e75-4911-a86e-29327b79b8f8\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.565058 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-config-data\") pod \"2842d110-9e75-4911-a86e-29327b79b8f8\" (UID: \"2842d110-9e75-4911-a86e-29327b79b8f8\") " Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.571092 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2842d110-9e75-4911-a86e-29327b79b8f8" (UID: "2842d110-9e75-4911-a86e-29327b79b8f8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.589402 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2842d110-9e75-4911-a86e-29327b79b8f8-kube-api-access-j67zx" (OuterVolumeSpecName: "kube-api-access-j67zx") pod "2842d110-9e75-4911-a86e-29327b79b8f8" (UID: "2842d110-9e75-4911-a86e-29327b79b8f8"). InnerVolumeSpecName "kube-api-access-j67zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.616801 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2842d110-9e75-4911-a86e-29327b79b8f8" (UID: "2842d110-9e75-4911-a86e-29327b79b8f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.639294 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-config-data" (OuterVolumeSpecName: "config-data") pod "2842d110-9e75-4911-a86e-29327b79b8f8" (UID: "2842d110-9e75-4911-a86e-29327b79b8f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.668522 4857 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.668909 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.668976 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j67zx\" (UniqueName: \"kubernetes.io/projected/2842d110-9e75-4911-a86e-29327b79b8f8-kube-api-access-j67zx\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.669001 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2842d110-9e75-4911-a86e-29327b79b8f8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.680920 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-r7zf9"] Dec 01 21:56:32 crc kubenswrapper[4857]: W1201 21:56:32.688764 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0906fcdc_a402_46cf_bf71_fcd7543a9034.slice/crio-113a14586ec2fb4384c8127a706c99a0c594eb62ec7067acc3eedaf6dbc413db WatchSource:0}: Error finding container 113a14586ec2fb4384c8127a706c99a0c594eb62ec7067acc3eedaf6dbc413db: Status 404 returned error can't find the container with id 113a14586ec2fb4384c8127a706c99a0c594eb62ec7067acc3eedaf6dbc413db Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.809476 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-r7zf9" event={"ID":"0906fcdc-a402-46cf-bf71-fcd7543a9034","Type":"ContainerStarted","Data":"113a14586ec2fb4384c8127a706c99a0c594eb62ec7067acc3eedaf6dbc413db"} Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.812139 4857 generic.go:334] "Generic (PLEG): container finished" podID="2842d110-9e75-4911-a86e-29327b79b8f8" containerID="621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b" exitCode=0 Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.812237 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-56454df68d-6lxzh" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.812370 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-56454df68d-6lxzh" event={"ID":"2842d110-9e75-4911-a86e-29327b79b8f8","Type":"ContainerDied","Data":"621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b"} Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.812475 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-56454df68d-6lxzh" event={"ID":"2842d110-9e75-4911-a86e-29327b79b8f8","Type":"ContainerDied","Data":"60c988026da607ec25e6c00365966a28d06b75048b35d258cee2d8a1e71cdb71"} Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.812561 4857 scope.go:117] "RemoveContainer" containerID="621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.850562 4857 scope.go:117] "RemoveContainer" containerID="621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b" Dec 01 21:56:32 crc kubenswrapper[4857]: E1201 21:56:32.851139 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b\": container with ID starting with 621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b not found: ID does not exist" containerID="621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.851246 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b"} err="failed to get container status \"621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b\": rpc error: code = NotFound desc = could not find container \"621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b\": container with ID starting with 621dac10a626292277c4e430a431593a8403ee2a79070c9af3819fa03d115f5b not found: ID does not exist" Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.874951 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-56454df68d-6lxzh"] Dec 01 21:56:32 crc kubenswrapper[4857]: I1201 21:56:32.883636 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-56454df68d-6lxzh"] Dec 01 21:56:33 crc kubenswrapper[4857]: I1201 21:56:33.483392 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 21:56:33 crc kubenswrapper[4857]: I1201 21:56:33.484324 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 21:56:33 crc kubenswrapper[4857]: I1201 21:56:33.518606 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 21:56:33 crc kubenswrapper[4857]: I1201 21:56:33.534750 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 21:56:33 crc kubenswrapper[4857]: I1201 21:56:33.851388 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2842d110-9e75-4911-a86e-29327b79b8f8" path="/var/lib/kubelet/pods/2842d110-9e75-4911-a86e-29327b79b8f8/volumes" Dec 01 21:56:33 crc kubenswrapper[4857]: I1201 21:56:33.852287 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 21:56:33 crc kubenswrapper[4857]: I1201 21:56:33.852306 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 21:56:35 crc kubenswrapper[4857]: I1201 21:56:35.877319 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:56:35 crc kubenswrapper[4857]: I1201 21:56:35.880826 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:56:36 crc kubenswrapper[4857]: I1201 21:56:36.216810 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 21:56:36 crc kubenswrapper[4857]: I1201 21:56:36.283959 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:36 crc kubenswrapper[4857]: I1201 21:56:36.284000 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:36 crc kubenswrapper[4857]: I1201 21:56:36.320235 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:36 crc kubenswrapper[4857]: I1201 21:56:36.375940 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 01 21:56:36 crc kubenswrapper[4857]: I1201 21:56:36.531415 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:36 crc kubenswrapper[4857]: I1201 21:56:36.888827 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:56:36 crc kubenswrapper[4857]: I1201 21:56:36.889123 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:36 crc kubenswrapper[4857]: I1201 21:56:36.890789 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:37 crc kubenswrapper[4857]: I1201 21:56:37.266134 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 21:56:38 crc kubenswrapper[4857]: I1201 21:56:38.523697 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 01 21:56:39 crc kubenswrapper[4857]: I1201 21:56:39.299768 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:39 crc kubenswrapper[4857]: I1201 21:56:39.299862 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 21:56:39 crc kubenswrapper[4857]: I1201 21:56:39.404938 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 21:56:48 crc kubenswrapper[4857]: I1201 21:56:48.030264 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-r7zf9" event={"ID":"0906fcdc-a402-46cf-bf71-fcd7543a9034","Type":"ContainerStarted","Data":"9e3539f6bb3a98cd815f508138d9ad0a369f92eba591b937717647a0ca512a72"} Dec 01 21:56:48 crc kubenswrapper[4857]: I1201 21:56:48.062590 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-r7zf9" podStartSLOduration=2.614920588 podStartE2EDuration="17.062561842s" podCreationTimestamp="2025-12-01 21:56:31 +0000 UTC" firstStartedPulling="2025-12-01 21:56:32.691953576 +0000 UTC m=+1311.182015893" lastFinishedPulling="2025-12-01 21:56:47.13959483 +0000 UTC m=+1325.629657147" observedRunningTime="2025-12-01 21:56:48.051940182 +0000 UTC m=+1326.542002489" watchObservedRunningTime="2025-12-01 21:56:48.062561842 +0000 UTC m=+1326.552624199" Dec 01 21:56:54 crc kubenswrapper[4857]: I1201 21:56:54.290642 4857 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod9dff5cc8-16db-4b2b-8d0a-72eb2c068c24"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod9dff5cc8-16db-4b2b-8d0a-72eb2c068c24] : Timed out while waiting for systemd to remove kubepods-besteffort-pod9dff5cc8_16db_4b2b_8d0a_72eb2c068c24.slice" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.135622 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.136778 4857 generic.go:334] "Generic (PLEG): container finished" podID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerID="6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440" exitCode=137 Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.136865 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913cd3df-6038-44e3-ad22-fc0fa6645058","Type":"ContainerDied","Data":"6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440"} Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.136947 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913cd3df-6038-44e3-ad22-fc0fa6645058","Type":"ContainerDied","Data":"04c300b0475348f55f4b71cccecccf3b2287989e898aab789a62953827605f01"} Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.136975 4857 scope.go:117] "RemoveContainer" containerID="6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.186932 4857 scope.go:117] "RemoveContainer" containerID="9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.226478 4857 scope.go:117] "RemoveContainer" containerID="014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.254296 4857 scope.go:117] "RemoveContainer" containerID="d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.284722 4857 scope.go:117] "RemoveContainer" containerID="6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440" Dec 01 21:56:57 crc kubenswrapper[4857]: E1201 21:56:57.285358 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440\": container with ID starting with 6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440 not found: ID does not exist" containerID="6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.285408 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440"} err="failed to get container status \"6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440\": rpc error: code = NotFound desc = could not find container \"6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440\": container with ID starting with 6f21cad33b9f6e7cbc9522366a7d74c0e59b207792bcfd7bcafdae5ec0c1f440 not found: ID does not exist" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.285447 4857 scope.go:117] "RemoveContainer" containerID="9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d" Dec 01 21:56:57 crc kubenswrapper[4857]: E1201 21:56:57.286189 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d\": container with ID starting with 9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d not found: ID does not exist" containerID="9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.286212 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d"} err="failed to get container status \"9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d\": rpc error: code = NotFound desc = could not find container \"9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d\": container with ID starting with 9752e541f6ab10fc93bce29514a5a24d911d7532b2bc07a8eb53ab6766749c3d not found: ID does not exist" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.286230 4857 scope.go:117] "RemoveContainer" containerID="014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c" Dec 01 21:56:57 crc kubenswrapper[4857]: E1201 21:56:57.286670 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c\": container with ID starting with 014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c not found: ID does not exist" containerID="014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.286710 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c"} err="failed to get container status \"014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c\": rpc error: code = NotFound desc = could not find container \"014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c\": container with ID starting with 014ab283815a35a7c798db17d5ff945bebf84cb885d9c95ea5c351faf975161c not found: ID does not exist" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.286742 4857 scope.go:117] "RemoveContainer" containerID="d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1" Dec 01 21:56:57 crc kubenswrapper[4857]: E1201 21:56:57.287096 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1\": container with ID starting with d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1 not found: ID does not exist" containerID="d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.287125 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1"} err="failed to get container status \"d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1\": rpc error: code = NotFound desc = could not find container \"d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1\": container with ID starting with d15963104af58caeb952dab0930041d09afe26696e48d97e4ea123955d7156c1 not found: ID does not exist" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.296704 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-sg-core-conf-yaml\") pod \"913cd3df-6038-44e3-ad22-fc0fa6645058\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.296743 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-combined-ca-bundle\") pod \"913cd3df-6038-44e3-ad22-fc0fa6645058\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.296834 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-config-data\") pod \"913cd3df-6038-44e3-ad22-fc0fa6645058\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.296883 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-scripts\") pod \"913cd3df-6038-44e3-ad22-fc0fa6645058\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.297062 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913cd3df-6038-44e3-ad22-fc0fa6645058-log-httpd\") pod \"913cd3df-6038-44e3-ad22-fc0fa6645058\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.297142 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913cd3df-6038-44e3-ad22-fc0fa6645058-run-httpd\") pod \"913cd3df-6038-44e3-ad22-fc0fa6645058\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.298068 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/913cd3df-6038-44e3-ad22-fc0fa6645058-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "913cd3df-6038-44e3-ad22-fc0fa6645058" (UID: "913cd3df-6038-44e3-ad22-fc0fa6645058"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.298355 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/913cd3df-6038-44e3-ad22-fc0fa6645058-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "913cd3df-6038-44e3-ad22-fc0fa6645058" (UID: "913cd3df-6038-44e3-ad22-fc0fa6645058"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.298436 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v586f\" (UniqueName: \"kubernetes.io/projected/913cd3df-6038-44e3-ad22-fc0fa6645058-kube-api-access-v586f\") pod \"913cd3df-6038-44e3-ad22-fc0fa6645058\" (UID: \"913cd3df-6038-44e3-ad22-fc0fa6645058\") " Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.299199 4857 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913cd3df-6038-44e3-ad22-fc0fa6645058-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.299218 4857 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913cd3df-6038-44e3-ad22-fc0fa6645058-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.304028 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-scripts" (OuterVolumeSpecName: "scripts") pod "913cd3df-6038-44e3-ad22-fc0fa6645058" (UID: "913cd3df-6038-44e3-ad22-fc0fa6645058"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.306422 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/913cd3df-6038-44e3-ad22-fc0fa6645058-kube-api-access-v586f" (OuterVolumeSpecName: "kube-api-access-v586f") pod "913cd3df-6038-44e3-ad22-fc0fa6645058" (UID: "913cd3df-6038-44e3-ad22-fc0fa6645058"). InnerVolumeSpecName "kube-api-access-v586f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.329513 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "913cd3df-6038-44e3-ad22-fc0fa6645058" (UID: "913cd3df-6038-44e3-ad22-fc0fa6645058"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.397732 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "913cd3df-6038-44e3-ad22-fc0fa6645058" (UID: "913cd3df-6038-44e3-ad22-fc0fa6645058"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.402334 4857 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.402372 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.402392 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.402410 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v586f\" (UniqueName: \"kubernetes.io/projected/913cd3df-6038-44e3-ad22-fc0fa6645058-kube-api-access-v586f\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.463602 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-config-data" (OuterVolumeSpecName: "config-data") pod "913cd3df-6038-44e3-ad22-fc0fa6645058" (UID: "913cd3df-6038-44e3-ad22-fc0fa6645058"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:56:57 crc kubenswrapper[4857]: I1201 21:56:57.503454 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913cd3df-6038-44e3-ad22-fc0fa6645058-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.147467 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.172542 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.182113 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.209952 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:56:58 crc kubenswrapper[4857]: E1201 21:56:58.210500 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2842d110-9e75-4911-a86e-29327b79b8f8" containerName="heat-engine" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.210524 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="2842d110-9e75-4911-a86e-29327b79b8f8" containerName="heat-engine" Dec 01 21:56:58 crc kubenswrapper[4857]: E1201 21:56:58.210558 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="proxy-httpd" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.210568 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="proxy-httpd" Dec 01 21:56:58 crc kubenswrapper[4857]: E1201 21:56:58.210598 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="sg-core" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.210606 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="sg-core" Dec 01 21:56:58 crc kubenswrapper[4857]: E1201 21:56:58.210624 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="ceilometer-notification-agent" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.210632 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="ceilometer-notification-agent" Dec 01 21:56:58 crc kubenswrapper[4857]: E1201 21:56:58.210645 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="ceilometer-central-agent" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.210655 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="ceilometer-central-agent" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.210866 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="2842d110-9e75-4911-a86e-29327b79b8f8" containerName="heat-engine" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.210888 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="sg-core" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.210910 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="ceilometer-central-agent" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.210923 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="proxy-httpd" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.210932 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" containerName="ceilometer-notification-agent" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.214905 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.218200 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.218240 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.234214 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.318059 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-config-data\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.318654 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/070fb164-a99c-481e-8e3d-4dbae13e2073-log-httpd\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.318712 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.318741 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.318766 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/070fb164-a99c-481e-8e3d-4dbae13e2073-run-httpd\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.319253 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxd2v\" (UniqueName: \"kubernetes.io/projected/070fb164-a99c-481e-8e3d-4dbae13e2073-kube-api-access-pxd2v\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.319401 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-scripts\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.421874 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxd2v\" (UniqueName: \"kubernetes.io/projected/070fb164-a99c-481e-8e3d-4dbae13e2073-kube-api-access-pxd2v\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.421938 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-scripts\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.421983 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-config-data\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.422048 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/070fb164-a99c-481e-8e3d-4dbae13e2073-log-httpd\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.422094 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.422143 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.422160 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/070fb164-a99c-481e-8e3d-4dbae13e2073-run-httpd\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.422678 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/070fb164-a99c-481e-8e3d-4dbae13e2073-log-httpd\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.422788 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/070fb164-a99c-481e-8e3d-4dbae13e2073-run-httpd\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.428617 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-scripts\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.430149 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-config-data\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.431712 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.434641 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.445342 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxd2v\" (UniqueName: \"kubernetes.io/projected/070fb164-a99c-481e-8e3d-4dbae13e2073-kube-api-access-pxd2v\") pod \"ceilometer-0\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " pod="openstack/ceilometer-0" Dec 01 21:56:58 crc kubenswrapper[4857]: I1201 21:56:58.535624 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:56:59 crc kubenswrapper[4857]: I1201 21:56:59.102743 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:56:59 crc kubenswrapper[4857]: W1201 21:56:59.107068 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod070fb164_a99c_481e_8e3d_4dbae13e2073.slice/crio-4e7c83e198682e0ffa8198690b339b9af52cc2abeb4e87465f16363243f30ccf WatchSource:0}: Error finding container 4e7c83e198682e0ffa8198690b339b9af52cc2abeb4e87465f16363243f30ccf: Status 404 returned error can't find the container with id 4e7c83e198682e0ffa8198690b339b9af52cc2abeb4e87465f16363243f30ccf Dec 01 21:56:59 crc kubenswrapper[4857]: I1201 21:56:59.111404 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 21:56:59 crc kubenswrapper[4857]: I1201 21:56:59.163700 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"070fb164-a99c-481e-8e3d-4dbae13e2073","Type":"ContainerStarted","Data":"4e7c83e198682e0ffa8198690b339b9af52cc2abeb4e87465f16363243f30ccf"} Dec 01 21:56:59 crc kubenswrapper[4857]: I1201 21:56:59.849110 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="913cd3df-6038-44e3-ad22-fc0fa6645058" path="/var/lib/kubelet/pods/913cd3df-6038-44e3-ad22-fc0fa6645058/volumes" Dec 01 21:57:00 crc kubenswrapper[4857]: I1201 21:57:00.176982 4857 generic.go:334] "Generic (PLEG): container finished" podID="0906fcdc-a402-46cf-bf71-fcd7543a9034" containerID="9e3539f6bb3a98cd815f508138d9ad0a369f92eba591b937717647a0ca512a72" exitCode=0 Dec 01 21:57:00 crc kubenswrapper[4857]: I1201 21:57:00.177057 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-r7zf9" event={"ID":"0906fcdc-a402-46cf-bf71-fcd7543a9034","Type":"ContainerDied","Data":"9e3539f6bb3a98cd815f508138d9ad0a369f92eba591b937717647a0ca512a72"} Dec 01 21:57:00 crc kubenswrapper[4857]: I1201 21:57:00.184881 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"070fb164-a99c-481e-8e3d-4dbae13e2073","Type":"ContainerStarted","Data":"70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116"} Dec 01 21:57:01 crc kubenswrapper[4857]: I1201 21:57:01.254805 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"070fb164-a99c-481e-8e3d-4dbae13e2073","Type":"ContainerStarted","Data":"f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4"} Dec 01 21:57:01 crc kubenswrapper[4857]: I1201 21:57:01.848781 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:57:01 crc kubenswrapper[4857]: I1201 21:57:01.905191 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-combined-ca-bundle\") pod \"0906fcdc-a402-46cf-bf71-fcd7543a9034\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " Dec 01 21:57:01 crc kubenswrapper[4857]: I1201 21:57:01.905293 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn6ht\" (UniqueName: \"kubernetes.io/projected/0906fcdc-a402-46cf-bf71-fcd7543a9034-kube-api-access-gn6ht\") pod \"0906fcdc-a402-46cf-bf71-fcd7543a9034\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " Dec 01 21:57:01 crc kubenswrapper[4857]: I1201 21:57:01.905333 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-config-data\") pod \"0906fcdc-a402-46cf-bf71-fcd7543a9034\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " Dec 01 21:57:01 crc kubenswrapper[4857]: I1201 21:57:01.905372 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-scripts\") pod \"0906fcdc-a402-46cf-bf71-fcd7543a9034\" (UID: \"0906fcdc-a402-46cf-bf71-fcd7543a9034\") " Dec 01 21:57:01 crc kubenswrapper[4857]: I1201 21:57:01.928008 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-scripts" (OuterVolumeSpecName: "scripts") pod "0906fcdc-a402-46cf-bf71-fcd7543a9034" (UID: "0906fcdc-a402-46cf-bf71-fcd7543a9034"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:01 crc kubenswrapper[4857]: I1201 21:57:01.932417 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0906fcdc-a402-46cf-bf71-fcd7543a9034-kube-api-access-gn6ht" (OuterVolumeSpecName: "kube-api-access-gn6ht") pod "0906fcdc-a402-46cf-bf71-fcd7543a9034" (UID: "0906fcdc-a402-46cf-bf71-fcd7543a9034"). InnerVolumeSpecName "kube-api-access-gn6ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:57:01 crc kubenswrapper[4857]: I1201 21:57:01.949548 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-config-data" (OuterVolumeSpecName: "config-data") pod "0906fcdc-a402-46cf-bf71-fcd7543a9034" (UID: "0906fcdc-a402-46cf-bf71-fcd7543a9034"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:01 crc kubenswrapper[4857]: I1201 21:57:01.998008 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0906fcdc-a402-46cf-bf71-fcd7543a9034" (UID: "0906fcdc-a402-46cf-bf71-fcd7543a9034"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.008497 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn6ht\" (UniqueName: \"kubernetes.io/projected/0906fcdc-a402-46cf-bf71-fcd7543a9034-kube-api-access-gn6ht\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.008535 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.008547 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.008560 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0906fcdc-a402-46cf-bf71-fcd7543a9034-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.269026 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-r7zf9" event={"ID":"0906fcdc-a402-46cf-bf71-fcd7543a9034","Type":"ContainerDied","Data":"113a14586ec2fb4384c8127a706c99a0c594eb62ec7067acc3eedaf6dbc413db"} Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.269078 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-r7zf9" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.269081 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="113a14586ec2fb4384c8127a706c99a0c594eb62ec7067acc3eedaf6dbc413db" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.276855 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"070fb164-a99c-481e-8e3d-4dbae13e2073","Type":"ContainerStarted","Data":"70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3"} Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.408434 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 21:57:02 crc kubenswrapper[4857]: E1201 21:57:02.409303 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0906fcdc-a402-46cf-bf71-fcd7543a9034" containerName="nova-cell0-conductor-db-sync" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.409326 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="0906fcdc-a402-46cf-bf71-fcd7543a9034" containerName="nova-cell0-conductor-db-sync" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.409559 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="0906fcdc-a402-46cf-bf71-fcd7543a9034" containerName="nova-cell0-conductor-db-sync" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.410384 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.412599 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fc42k" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.413417 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.418447 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.517835 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a69a1083-f915-4820-82f3-f82c5053d365-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a69a1083-f915-4820-82f3-f82c5053d365\") " pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.518477 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a69a1083-f915-4820-82f3-f82c5053d365-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a69a1083-f915-4820-82f3-f82c5053d365\") " pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.518617 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2ww7\" (UniqueName: \"kubernetes.io/projected/a69a1083-f915-4820-82f3-f82c5053d365-kube-api-access-w2ww7\") pod \"nova-cell0-conductor-0\" (UID: \"a69a1083-f915-4820-82f3-f82c5053d365\") " pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.623324 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a69a1083-f915-4820-82f3-f82c5053d365-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a69a1083-f915-4820-82f3-f82c5053d365\") " pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.623829 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2ww7\" (UniqueName: \"kubernetes.io/projected/a69a1083-f915-4820-82f3-f82c5053d365-kube-api-access-w2ww7\") pod \"nova-cell0-conductor-0\" (UID: \"a69a1083-f915-4820-82f3-f82c5053d365\") " pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.624107 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a69a1083-f915-4820-82f3-f82c5053d365-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a69a1083-f915-4820-82f3-f82c5053d365\") " pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.630073 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a69a1083-f915-4820-82f3-f82c5053d365-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a69a1083-f915-4820-82f3-f82c5053d365\") " pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.630560 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a69a1083-f915-4820-82f3-f82c5053d365-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a69a1083-f915-4820-82f3-f82c5053d365\") " pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.644250 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2ww7\" (UniqueName: \"kubernetes.io/projected/a69a1083-f915-4820-82f3-f82c5053d365-kube-api-access-w2ww7\") pod \"nova-cell0-conductor-0\" (UID: \"a69a1083-f915-4820-82f3-f82c5053d365\") " pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:02 crc kubenswrapper[4857]: I1201 21:57:02.732476 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:03 crc kubenswrapper[4857]: I1201 21:57:03.320614 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 21:57:03 crc kubenswrapper[4857]: W1201 21:57:03.325401 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda69a1083_f915_4820_82f3_f82c5053d365.slice/crio-92c5c0f291e658ba85a532edcd9899a967219f1e6f5138471082c096b7102280 WatchSource:0}: Error finding container 92c5c0f291e658ba85a532edcd9899a967219f1e6f5138471082c096b7102280: Status 404 returned error can't find the container with id 92c5c0f291e658ba85a532edcd9899a967219f1e6f5138471082c096b7102280 Dec 01 21:57:04 crc kubenswrapper[4857]: I1201 21:57:04.297965 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a69a1083-f915-4820-82f3-f82c5053d365","Type":"ContainerStarted","Data":"65c27a8f268c7d1d56d2e889378f262195031fe1fdb3fdfddc0ad4a670279e4a"} Dec 01 21:57:04 crc kubenswrapper[4857]: I1201 21:57:04.298732 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a69a1083-f915-4820-82f3-f82c5053d365","Type":"ContainerStarted","Data":"92c5c0f291e658ba85a532edcd9899a967219f1e6f5138471082c096b7102280"} Dec 01 21:57:04 crc kubenswrapper[4857]: I1201 21:57:04.298760 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:04 crc kubenswrapper[4857]: I1201 21:57:04.302903 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"070fb164-a99c-481e-8e3d-4dbae13e2073","Type":"ContainerStarted","Data":"046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c"} Dec 01 21:57:04 crc kubenswrapper[4857]: I1201 21:57:04.303120 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 21:57:04 crc kubenswrapper[4857]: I1201 21:57:04.322151 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.322009557 podStartE2EDuration="2.322009557s" podCreationTimestamp="2025-12-01 21:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:57:04.316033321 +0000 UTC m=+1342.806095658" watchObservedRunningTime="2025-12-01 21:57:04.322009557 +0000 UTC m=+1342.812071884" Dec 01 21:57:04 crc kubenswrapper[4857]: I1201 21:57:04.362278 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.705215972 podStartE2EDuration="6.362259874s" podCreationTimestamp="2025-12-01 21:56:58 +0000 UTC" firstStartedPulling="2025-12-01 21:56:59.111179161 +0000 UTC m=+1337.601241478" lastFinishedPulling="2025-12-01 21:57:03.768223063 +0000 UTC m=+1342.258285380" observedRunningTime="2025-12-01 21:57:04.357593129 +0000 UTC m=+1342.847655466" watchObservedRunningTime="2025-12-01 21:57:04.362259874 +0000 UTC m=+1342.852322191" Dec 01 21:57:06 crc kubenswrapper[4857]: I1201 21:57:06.587654 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:57:06 crc kubenswrapper[4857]: I1201 21:57:06.588302 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="ceilometer-central-agent" containerID="cri-o://70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116" gracePeriod=30 Dec 01 21:57:06 crc kubenswrapper[4857]: I1201 21:57:06.588385 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="ceilometer-notification-agent" containerID="cri-o://f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4" gracePeriod=30 Dec 01 21:57:06 crc kubenswrapper[4857]: I1201 21:57:06.588385 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="sg-core" containerID="cri-o://70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3" gracePeriod=30 Dec 01 21:57:06 crc kubenswrapper[4857]: I1201 21:57:06.588558 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="proxy-httpd" containerID="cri-o://046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c" gracePeriod=30 Dec 01 21:57:07 crc kubenswrapper[4857]: I1201 21:57:07.343585 4857 generic.go:334] "Generic (PLEG): container finished" podID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerID="046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c" exitCode=0 Dec 01 21:57:07 crc kubenswrapper[4857]: I1201 21:57:07.343638 4857 generic.go:334] "Generic (PLEG): container finished" podID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerID="70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3" exitCode=2 Dec 01 21:57:07 crc kubenswrapper[4857]: I1201 21:57:07.343654 4857 generic.go:334] "Generic (PLEG): container finished" podID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerID="f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4" exitCode=0 Dec 01 21:57:07 crc kubenswrapper[4857]: I1201 21:57:07.343681 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"070fb164-a99c-481e-8e3d-4dbae13e2073","Type":"ContainerDied","Data":"046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c"} Dec 01 21:57:07 crc kubenswrapper[4857]: I1201 21:57:07.343727 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"070fb164-a99c-481e-8e3d-4dbae13e2073","Type":"ContainerDied","Data":"70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3"} Dec 01 21:57:07 crc kubenswrapper[4857]: I1201 21:57:07.343740 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"070fb164-a99c-481e-8e3d-4dbae13e2073","Type":"ContainerDied","Data":"f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4"} Dec 01 21:57:12 crc kubenswrapper[4857]: I1201 21:57:12.769537 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.148443 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.257011 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/070fb164-a99c-481e-8e3d-4dbae13e2073-run-httpd\") pod \"070fb164-a99c-481e-8e3d-4dbae13e2073\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.257077 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-scripts\") pod \"070fb164-a99c-481e-8e3d-4dbae13e2073\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.257172 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-config-data\") pod \"070fb164-a99c-481e-8e3d-4dbae13e2073\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.257195 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-sg-core-conf-yaml\") pod \"070fb164-a99c-481e-8e3d-4dbae13e2073\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.257237 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-combined-ca-bundle\") pod \"070fb164-a99c-481e-8e3d-4dbae13e2073\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.257321 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxd2v\" (UniqueName: \"kubernetes.io/projected/070fb164-a99c-481e-8e3d-4dbae13e2073-kube-api-access-pxd2v\") pod \"070fb164-a99c-481e-8e3d-4dbae13e2073\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.257346 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/070fb164-a99c-481e-8e3d-4dbae13e2073-log-httpd\") pod \"070fb164-a99c-481e-8e3d-4dbae13e2073\" (UID: \"070fb164-a99c-481e-8e3d-4dbae13e2073\") " Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.257569 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/070fb164-a99c-481e-8e3d-4dbae13e2073-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "070fb164-a99c-481e-8e3d-4dbae13e2073" (UID: "070fb164-a99c-481e-8e3d-4dbae13e2073"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.257908 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/070fb164-a99c-481e-8e3d-4dbae13e2073-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "070fb164-a99c-481e-8e3d-4dbae13e2073" (UID: "070fb164-a99c-481e-8e3d-4dbae13e2073"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.258391 4857 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/070fb164-a99c-481e-8e3d-4dbae13e2073-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.258421 4857 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/070fb164-a99c-481e-8e3d-4dbae13e2073-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.266226 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-scripts" (OuterVolumeSpecName: "scripts") pod "070fb164-a99c-481e-8e3d-4dbae13e2073" (UID: "070fb164-a99c-481e-8e3d-4dbae13e2073"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.266299 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/070fb164-a99c-481e-8e3d-4dbae13e2073-kube-api-access-pxd2v" (OuterVolumeSpecName: "kube-api-access-pxd2v") pod "070fb164-a99c-481e-8e3d-4dbae13e2073" (UID: "070fb164-a99c-481e-8e3d-4dbae13e2073"). InnerVolumeSpecName "kube-api-access-pxd2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.296583 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-m4rd7"] Dec 01 21:57:13 crc kubenswrapper[4857]: E1201 21:57:13.297117 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="ceilometer-notification-agent" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.297140 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="ceilometer-notification-agent" Dec 01 21:57:13 crc kubenswrapper[4857]: E1201 21:57:13.297181 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="ceilometer-central-agent" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.297189 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="ceilometer-central-agent" Dec 01 21:57:13 crc kubenswrapper[4857]: E1201 21:57:13.297210 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="proxy-httpd" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.297218 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="proxy-httpd" Dec 01 21:57:13 crc kubenswrapper[4857]: E1201 21:57:13.297239 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="sg-core" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.297246 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="sg-core" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.297470 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="ceilometer-central-agent" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.297496 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="sg-core" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.297529 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="proxy-httpd" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.297545 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerName="ceilometer-notification-agent" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.298678 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.300978 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.301682 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.311695 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "070fb164-a99c-481e-8e3d-4dbae13e2073" (UID: "070fb164-a99c-481e-8e3d-4dbae13e2073"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.337591 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-m4rd7"] Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.359858 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.360201 4857 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.360211 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxd2v\" (UniqueName: \"kubernetes.io/projected/070fb164-a99c-481e-8e3d-4dbae13e2073-kube-api-access-pxd2v\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.419975 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "070fb164-a99c-481e-8e3d-4dbae13e2073" (UID: "070fb164-a99c-481e-8e3d-4dbae13e2073"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.463319 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-config-data\") pod \"nova-cell0-cell-mapping-m4rd7\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.463502 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-scripts\") pod \"nova-cell0-cell-mapping-m4rd7\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.463569 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-m4rd7\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.463592 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpl7q\" (UniqueName: \"kubernetes.io/projected/70f52857-7d57-421e-8bf0-b849a8c02ffe-kube-api-access-gpl7q\") pod \"nova-cell0-cell-mapping-m4rd7\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.463657 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.464086 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.465728 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.476071 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.476278 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.481009 4857 generic.go:334] "Generic (PLEG): container finished" podID="070fb164-a99c-481e-8e3d-4dbae13e2073" containerID="70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116" exitCode=0 Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.481079 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"070fb164-a99c-481e-8e3d-4dbae13e2073","Type":"ContainerDied","Data":"70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116"} Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.481116 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"070fb164-a99c-481e-8e3d-4dbae13e2073","Type":"ContainerDied","Data":"4e7c83e198682e0ffa8198690b339b9af52cc2abeb4e87465f16363243f30ccf"} Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.481141 4857 scope.go:117] "RemoveContainer" containerID="046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.481380 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.549772 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-config-data" (OuterVolumeSpecName: "config-data") pod "070fb164-a99c-481e-8e3d-4dbae13e2073" (UID: "070fb164-a99c-481e-8e3d-4dbae13e2073"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.564151 4857 scope.go:117] "RemoveContainer" containerID="70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.567678 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-m4rd7\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.567742 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpl7q\" (UniqueName: \"kubernetes.io/projected/70f52857-7d57-421e-8bf0-b849a8c02ffe-kube-api-access-gpl7q\") pod \"nova-cell0-cell-mapping-m4rd7\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.567777 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-config-data\") pod \"nova-scheduler-0\" (UID: \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.567800 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.567834 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-config-data\") pod \"nova-cell0-cell-mapping-m4rd7\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.567953 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svmlg\" (UniqueName: \"kubernetes.io/projected/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-kube-api-access-svmlg\") pod \"nova-scheduler-0\" (UID: \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.568012 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-scripts\") pod \"nova-cell0-cell-mapping-m4rd7\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.568111 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070fb164-a99c-481e-8e3d-4dbae13e2073-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.572509 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-m4rd7\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.579452 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-scripts\") pod \"nova-cell0-cell-mapping-m4rd7\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.581702 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-config-data\") pod \"nova-cell0-cell-mapping-m4rd7\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.596661 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpl7q\" (UniqueName: \"kubernetes.io/projected/70f52857-7d57-421e-8bf0-b849a8c02ffe-kube-api-access-gpl7q\") pod \"nova-cell0-cell-mapping-m4rd7\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.623034 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.628796 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.636478 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.647895 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.658809 4857 scope.go:117] "RemoveContainer" containerID="f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.675834 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-config-data\") pod \"nova-scheduler-0\" (UID: \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.675868 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.675968 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svmlg\" (UniqueName: \"kubernetes.io/projected/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-kube-api-access-svmlg\") pod \"nova-scheduler-0\" (UID: \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.688590 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-config-data\") pod \"nova-scheduler-0\" (UID: \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.688791 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.707419 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.709654 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.730073 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.730182 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svmlg\" (UniqueName: \"kubernetes.io/projected/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-kube-api-access-svmlg\") pod \"nova-scheduler-0\" (UID: \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.741681 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.774251 4857 scope.go:117] "RemoveContainer" containerID="70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.788765 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9e426f5-5fde-4067-8128-76f26f907fc8-logs\") pod \"nova-api-0\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.788896 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-logs\") pod \"nova-metadata-0\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.788929 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.789168 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l4vh\" (UniqueName: \"kubernetes.io/projected/c9e426f5-5fde-4067-8128-76f26f907fc8-kube-api-access-4l4vh\") pod \"nova-api-0\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.789327 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e426f5-5fde-4067-8128-76f26f907fc8-config-data\") pod \"nova-api-0\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.789371 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z5ls\" (UniqueName: \"kubernetes.io/projected/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-kube-api-access-7z5ls\") pod \"nova-metadata-0\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.789530 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-config-data\") pod \"nova-metadata-0\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.789551 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e426f5-5fde-4067-8128-76f26f907fc8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.790959 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.804599 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.827698 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-fl6ws"] Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.829405 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.890856 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-fl6ws"] Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.891261 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.892107 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-config-data\") pod \"nova-metadata-0\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.892544 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e426f5-5fde-4067-8128-76f26f907fc8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.893107 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.893440 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.893576 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9e426f5-5fde-4067-8128-76f26f907fc8-logs\") pod \"nova-api-0\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.893828 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-logs\") pod \"nova-metadata-0\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.893854 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.893903 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9e426f5-5fde-4067-8128-76f26f907fc8-logs\") pod \"nova-api-0\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.893952 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l4vh\" (UniqueName: \"kubernetes.io/projected/c9e426f5-5fde-4067-8128-76f26f907fc8-kube-api-access-4l4vh\") pod \"nova-api-0\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.894023 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e426f5-5fde-4067-8128-76f26f907fc8-config-data\") pod \"nova-api-0\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.894066 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z5ls\" (UniqueName: \"kubernetes.io/projected/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-kube-api-access-7z5ls\") pod \"nova-metadata-0\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.894317 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-logs\") pod \"nova-metadata-0\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.936889 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-config-data\") pod \"nova-metadata-0\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.937525 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e426f5-5fde-4067-8128-76f26f907fc8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.937919 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.938330 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e426f5-5fde-4067-8128-76f26f907fc8-config-data\") pod \"nova-api-0\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.938979 4857 scope.go:117] "RemoveContainer" containerID="046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c" Dec 01 21:57:13 crc kubenswrapper[4857]: E1201 21:57:13.946802 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c\": container with ID starting with 046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c not found: ID does not exist" containerID="046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.947190 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c"} err="failed to get container status \"046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c\": rpc error: code = NotFound desc = could not find container \"046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c\": container with ID starting with 046746fbd23982330e7b858ac99e70388cb3061255ddddcd4c3c9fc655200a3c not found: ID does not exist" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.947298 4857 scope.go:117] "RemoveContainer" containerID="70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3" Dec 01 21:57:13 crc kubenswrapper[4857]: E1201 21:57:13.949352 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3\": container with ID starting with 70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3 not found: ID does not exist" containerID="70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.949411 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3"} err="failed to get container status \"70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3\": rpc error: code = NotFound desc = could not find container \"70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3\": container with ID starting with 70b21f02a0734f197685d633badea2222c12e685812455ac09a03881083323c3 not found: ID does not exist" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.949447 4857 scope.go:117] "RemoveContainer" containerID="f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4" Dec 01 21:57:13 crc kubenswrapper[4857]: E1201 21:57:13.950946 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4\": container with ID starting with f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4 not found: ID does not exist" containerID="f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.951137 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4"} err="failed to get container status \"f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4\": rpc error: code = NotFound desc = could not find container \"f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4\": container with ID starting with f518b8853bb5ae348da3a94b60c8c82df96f47e59fbe69f93409258942696ee4 not found: ID does not exist" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.951241 4857 scope.go:117] "RemoveContainer" containerID="70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116" Dec 01 21:57:13 crc kubenswrapper[4857]: E1201 21:57:13.952524 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116\": container with ID starting with 70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116 not found: ID does not exist" containerID="70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.952648 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116"} err="failed to get container status \"70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116\": rpc error: code = NotFound desc = could not find container \"70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116\": container with ID starting with 70e06567399918b373a7bceb0419687cb0821ded5f64cdcbc67771d421795116 not found: ID does not exist" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.961200 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.974593 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z5ls\" (UniqueName: \"kubernetes.io/projected/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-kube-api-access-7z5ls\") pod \"nova-metadata-0\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " pod="openstack/nova-metadata-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.977595 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l4vh\" (UniqueName: \"kubernetes.io/projected/c9e426f5-5fde-4067-8128-76f26f907fc8-kube-api-access-4l4vh\") pod \"nova-api-0\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " pod="openstack/nova-api-0" Dec 01 21:57:13 crc kubenswrapper[4857]: I1201 21:57:13.991389 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.003227 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwsr2\" (UniqueName: \"kubernetes.io/projected/fc6ab107-b310-4f97-8bef-695b1a75f728-kube-api-access-mwsr2\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.003315 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.003388 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-dns-svc\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.003574 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562073c1-2475-4e08-a121-b69cf751fa63-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"562073c1-2475-4e08-a121-b69cf751fa63\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.009108 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz9z2\" (UniqueName: \"kubernetes.io/projected/562073c1-2475-4e08-a121-b69cf751fa63-kube-api-access-vz9z2\") pod \"nova-cell1-novncproxy-0\" (UID: \"562073c1-2475-4e08-a121-b69cf751fa63\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.009238 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.009291 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-config\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.009431 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.009484 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/562073c1-2475-4e08-a121-b69cf751fa63-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"562073c1-2475-4e08-a121-b69cf751fa63\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.168162 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.168234 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-dns-svc\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.168444 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562073c1-2475-4e08-a121-b69cf751fa63-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"562073c1-2475-4e08-a121-b69cf751fa63\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.168601 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz9z2\" (UniqueName: \"kubernetes.io/projected/562073c1-2475-4e08-a121-b69cf751fa63-kube-api-access-vz9z2\") pod \"nova-cell1-novncproxy-0\" (UID: \"562073c1-2475-4e08-a121-b69cf751fa63\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.168665 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.168695 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-config\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.168773 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.168807 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/562073c1-2475-4e08-a121-b69cf751fa63-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"562073c1-2475-4e08-a121-b69cf751fa63\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.168884 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwsr2\" (UniqueName: \"kubernetes.io/projected/fc6ab107-b310-4f97-8bef-695b1a75f728-kube-api-access-mwsr2\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.171572 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.172435 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.174058 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-config\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.174221 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-dns-svc\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.185006 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562073c1-2475-4e08-a121-b69cf751fa63-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"562073c1-2475-4e08-a121-b69cf751fa63\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.188763 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.205486 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/562073c1-2475-4e08-a121-b69cf751fa63-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"562073c1-2475-4e08-a121-b69cf751fa63\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.214175 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz9z2\" (UniqueName: \"kubernetes.io/projected/562073c1-2475-4e08-a121-b69cf751fa63-kube-api-access-vz9z2\") pod \"nova-cell1-novncproxy-0\" (UID: \"562073c1-2475-4e08-a121-b69cf751fa63\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.226106 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwsr2\" (UniqueName: \"kubernetes.io/projected/fc6ab107-b310-4f97-8bef-695b1a75f728-kube-api-access-mwsr2\") pod \"dnsmasq-dns-9b86998b5-fl6ws\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.270447 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.288309 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.338965 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.756774 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-m4rd7"] Dec 01 21:57:14 crc kubenswrapper[4857]: W1201 21:57:14.769305 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70f52857_7d57_421e_8bf0_b849a8c02ffe.slice/crio-530f8de5bd12e9b9d3b0d768c71ab38f16d8997ec05bd4d1c26a35cd827a6e30 WatchSource:0}: Error finding container 530f8de5bd12e9b9d3b0d768c71ab38f16d8997ec05bd4d1c26a35cd827a6e30: Status 404 returned error can't find the container with id 530f8de5bd12e9b9d3b0d768c71ab38f16d8997ec05bd4d1c26a35cd827a6e30 Dec 01 21:57:14 crc kubenswrapper[4857]: I1201 21:57:14.926971 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.021019 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.093181 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-v9qs5"] Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.094772 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.098591 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.098868 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.118775 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-v9qs5"] Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.130222 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-fl6ws"] Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.140866 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.200618 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-v9qs5\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.200718 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zffkc\" (UniqueName: \"kubernetes.io/projected/9df6644a-8870-44de-8e78-51f76e062f85-kube-api-access-zffkc\") pod \"nova-cell1-conductor-db-sync-v9qs5\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.200838 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-scripts\") pod \"nova-cell1-conductor-db-sync-v9qs5\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.200874 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-config-data\") pod \"nova-cell1-conductor-db-sync-v9qs5\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.220527 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.302198 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zffkc\" (UniqueName: \"kubernetes.io/projected/9df6644a-8870-44de-8e78-51f76e062f85-kube-api-access-zffkc\") pod \"nova-cell1-conductor-db-sync-v9qs5\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.302310 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-scripts\") pod \"nova-cell1-conductor-db-sync-v9qs5\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.302330 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-config-data\") pod \"nova-cell1-conductor-db-sync-v9qs5\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.302404 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-v9qs5\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.307016 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-v9qs5\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.307403 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-scripts\") pod \"nova-cell1-conductor-db-sync-v9qs5\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.308134 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-config-data\") pod \"nova-cell1-conductor-db-sync-v9qs5\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.326849 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zffkc\" (UniqueName: \"kubernetes.io/projected/9df6644a-8870-44de-8e78-51f76e062f85-kube-api-access-zffkc\") pod \"nova-cell1-conductor-db-sync-v9qs5\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.422689 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.528371 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"562073c1-2475-4e08-a121-b69cf751fa63","Type":"ContainerStarted","Data":"e87be741b0a475c9cf4b961d4cd0938e07dfa19c74eedfcc73dc2da97429867d"} Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.532874 4857 generic.go:334] "Generic (PLEG): container finished" podID="fc6ab107-b310-4f97-8bef-695b1a75f728" containerID="e4fce0baf3b19bf0c6fd344c73551228cb88e80e1298c2b019f82171e1b3d532" exitCode=0 Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.532940 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" event={"ID":"fc6ab107-b310-4f97-8bef-695b1a75f728","Type":"ContainerDied","Data":"e4fce0baf3b19bf0c6fd344c73551228cb88e80e1298c2b019f82171e1b3d532"} Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.532961 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" event={"ID":"fc6ab107-b310-4f97-8bef-695b1a75f728","Type":"ContainerStarted","Data":"27e5949444e915d4996447210ab491a80f406be7138117caf1f2fd73c52fc255"} Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.538324 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"88c65caf-85af-4b62-bcb4-5980ab0ef3dc","Type":"ContainerStarted","Data":"fe5502aeb26e008e03231242e0f68c3678fe0d7d1cc0829d9a894c9d5902f2d5"} Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.541059 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b","Type":"ContainerStarted","Data":"7a9cdce76094fedcccb39a4ffc860de9c3f53a398bfca8c494393844b65d8a3b"} Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.586642 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-m4rd7" event={"ID":"70f52857-7d57-421e-8bf0-b849a8c02ffe","Type":"ContainerStarted","Data":"f64fce28262cd259de4d3107ccf2fd6b7a1b3c2c8b4babaae8b518b5223d8aba"} Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.586717 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-m4rd7" event={"ID":"70f52857-7d57-421e-8bf0-b849a8c02ffe","Type":"ContainerStarted","Data":"530f8de5bd12e9b9d3b0d768c71ab38f16d8997ec05bd4d1c26a35cd827a6e30"} Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.595263 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c9e426f5-5fde-4067-8128-76f26f907fc8","Type":"ContainerStarted","Data":"dc1ca86002a0f1e47ab5646e6827a820f0eaa87a31454c2253efda60cb2f29bf"} Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.660211 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-m4rd7" podStartSLOduration=2.6601888320000002 podStartE2EDuration="2.660188832s" podCreationTimestamp="2025-12-01 21:57:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:57:15.646017874 +0000 UTC m=+1354.136080191" watchObservedRunningTime="2025-12-01 21:57:15.660188832 +0000 UTC m=+1354.150251149" Dec 01 21:57:15 crc kubenswrapper[4857]: I1201 21:57:15.981418 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-v9qs5"] Dec 01 21:57:16 crc kubenswrapper[4857]: I1201 21:57:16.618490 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" event={"ID":"fc6ab107-b310-4f97-8bef-695b1a75f728","Type":"ContainerStarted","Data":"c5fe7d8f3d0454204f851dc543dcccd87f04767356e8108a8c524c3a66dec83b"} Dec 01 21:57:16 crc kubenswrapper[4857]: I1201 21:57:16.619735 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:16 crc kubenswrapper[4857]: I1201 21:57:16.621384 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-v9qs5" event={"ID":"9df6644a-8870-44de-8e78-51f76e062f85","Type":"ContainerStarted","Data":"d78b0cfe9bd8d514b110c57c6cac5a8599248656c2b26a32beb8199cf1eae8c6"} Dec 01 21:57:16 crc kubenswrapper[4857]: I1201 21:57:16.621439 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-v9qs5" event={"ID":"9df6644a-8870-44de-8e78-51f76e062f85","Type":"ContainerStarted","Data":"739fc919a4043e77d470a4bc525f5f5bc673415c029400f17fdcf3a09aaff914"} Dec 01 21:57:16 crc kubenswrapper[4857]: I1201 21:57:16.641977 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" podStartSLOduration=3.641957776 podStartE2EDuration="3.641957776s" podCreationTimestamp="2025-12-01 21:57:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:57:16.64009371 +0000 UTC m=+1355.130156047" watchObservedRunningTime="2025-12-01 21:57:16.641957776 +0000 UTC m=+1355.132020093" Dec 01 21:57:16 crc kubenswrapper[4857]: I1201 21:57:16.659168 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-v9qs5" podStartSLOduration=1.6591353070000001 podStartE2EDuration="1.659135307s" podCreationTimestamp="2025-12-01 21:57:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:57:16.656216555 +0000 UTC m=+1355.146278872" watchObservedRunningTime="2025-12-01 21:57:16.659135307 +0000 UTC m=+1355.149197634" Dec 01 21:57:18 crc kubenswrapper[4857]: I1201 21:57:18.400454 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:18 crc kubenswrapper[4857]: I1201 21:57:18.415805 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.708255 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"562073c1-2475-4e08-a121-b69cf751fa63","Type":"ContainerStarted","Data":"202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce"} Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.709267 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="562073c1-2475-4e08-a121-b69cf751fa63" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce" gracePeriod=30 Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.718123 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"88c65caf-85af-4b62-bcb4-5980ab0ef3dc","Type":"ContainerStarted","Data":"c4c0303796540659ca714ede5624b707792775bc31659e86472eee7d7f67d243"} Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.718193 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"88c65caf-85af-4b62-bcb4-5980ab0ef3dc","Type":"ContainerStarted","Data":"e1eb7b2324f2b07c77d704df8935422c1179839daba017b69e5921a6b452affd"} Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.718371 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="88c65caf-85af-4b62-bcb4-5980ab0ef3dc" containerName="nova-metadata-log" containerID="cri-o://e1eb7b2324f2b07c77d704df8935422c1179839daba017b69e5921a6b452affd" gracePeriod=30 Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.718538 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="88c65caf-85af-4b62-bcb4-5980ab0ef3dc" containerName="nova-metadata-metadata" containerID="cri-o://c4c0303796540659ca714ede5624b707792775bc31659e86472eee7d7f67d243" gracePeriod=30 Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.724789 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b","Type":"ContainerStarted","Data":"d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c"} Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.733302 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c9e426f5-5fde-4067-8128-76f26f907fc8","Type":"ContainerStarted","Data":"d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752"} Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.733382 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c9e426f5-5fde-4067-8128-76f26f907fc8","Type":"ContainerStarted","Data":"12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c"} Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.758021 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.409753545 podStartE2EDuration="8.757991578s" podCreationTimestamp="2025-12-01 21:57:13 +0000 UTC" firstStartedPulling="2025-12-01 21:57:15.227809743 +0000 UTC m=+1353.717872060" lastFinishedPulling="2025-12-01 21:57:20.576047776 +0000 UTC m=+1359.066110093" observedRunningTime="2025-12-01 21:57:21.735456495 +0000 UTC m=+1360.225518822" watchObservedRunningTime="2025-12-01 21:57:21.757991578 +0000 UTC m=+1360.248053905" Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.765971 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.205306224 podStartE2EDuration="8.765938823s" podCreationTimestamp="2025-12-01 21:57:13 +0000 UTC" firstStartedPulling="2025-12-01 21:57:15.027273268 +0000 UTC m=+1353.517335585" lastFinishedPulling="2025-12-01 21:57:20.587905867 +0000 UTC m=+1359.077968184" observedRunningTime="2025-12-01 21:57:21.75849233 +0000 UTC m=+1360.248554647" watchObservedRunningTime="2025-12-01 21:57:21.765938823 +0000 UTC m=+1360.256001140" Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.789958 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.160097235 podStartE2EDuration="8.789930731s" podCreationTimestamp="2025-12-01 21:57:13 +0000 UTC" firstStartedPulling="2025-12-01 21:57:14.94619351 +0000 UTC m=+1353.436255827" lastFinishedPulling="2025-12-01 21:57:20.576027006 +0000 UTC m=+1359.066089323" observedRunningTime="2025-12-01 21:57:21.779949466 +0000 UTC m=+1360.270011783" watchObservedRunningTime="2025-12-01 21:57:21.789930731 +0000 UTC m=+1360.279993058" Dec 01 21:57:21 crc kubenswrapper[4857]: I1201 21:57:21.804931 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.366094405 podStartE2EDuration="8.804903418s" podCreationTimestamp="2025-12-01 21:57:13 +0000 UTC" firstStartedPulling="2025-12-01 21:57:15.13832505 +0000 UTC m=+1353.628387367" lastFinishedPulling="2025-12-01 21:57:20.577134063 +0000 UTC m=+1359.067196380" observedRunningTime="2025-12-01 21:57:21.798572153 +0000 UTC m=+1360.288634470" watchObservedRunningTime="2025-12-01 21:57:21.804903418 +0000 UTC m=+1360.294965745" Dec 01 21:57:22 crc kubenswrapper[4857]: I1201 21:57:22.752454 4857 generic.go:334] "Generic (PLEG): container finished" podID="88c65caf-85af-4b62-bcb4-5980ab0ef3dc" containerID="c4c0303796540659ca714ede5624b707792775bc31659e86472eee7d7f67d243" exitCode=0 Dec 01 21:57:22 crc kubenswrapper[4857]: I1201 21:57:22.752936 4857 generic.go:334] "Generic (PLEG): container finished" podID="88c65caf-85af-4b62-bcb4-5980ab0ef3dc" containerID="e1eb7b2324f2b07c77d704df8935422c1179839daba017b69e5921a6b452affd" exitCode=143 Dec 01 21:57:22 crc kubenswrapper[4857]: I1201 21:57:22.752534 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"88c65caf-85af-4b62-bcb4-5980ab0ef3dc","Type":"ContainerDied","Data":"c4c0303796540659ca714ede5624b707792775bc31659e86472eee7d7f67d243"} Dec 01 21:57:22 crc kubenswrapper[4857]: I1201 21:57:22.753096 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"88c65caf-85af-4b62-bcb4-5980ab0ef3dc","Type":"ContainerDied","Data":"e1eb7b2324f2b07c77d704df8935422c1179839daba017b69e5921a6b452affd"} Dec 01 21:57:22 crc kubenswrapper[4857]: I1201 21:57:22.866307 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.015133 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-logs\") pod \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.015348 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-combined-ca-bundle\") pod \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.015488 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-config-data\") pod \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.017641 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7z5ls\" (UniqueName: \"kubernetes.io/projected/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-kube-api-access-7z5ls\") pod \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\" (UID: \"88c65caf-85af-4b62-bcb4-5980ab0ef3dc\") " Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.021030 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-logs" (OuterVolumeSpecName: "logs") pod "88c65caf-85af-4b62-bcb4-5980ab0ef3dc" (UID: "88c65caf-85af-4b62-bcb4-5980ab0ef3dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.030303 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-kube-api-access-7z5ls" (OuterVolumeSpecName: "kube-api-access-7z5ls") pod "88c65caf-85af-4b62-bcb4-5980ab0ef3dc" (UID: "88c65caf-85af-4b62-bcb4-5980ab0ef3dc"). InnerVolumeSpecName "kube-api-access-7z5ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.062443 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-config-data" (OuterVolumeSpecName: "config-data") pod "88c65caf-85af-4b62-bcb4-5980ab0ef3dc" (UID: "88c65caf-85af-4b62-bcb4-5980ab0ef3dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.073352 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88c65caf-85af-4b62-bcb4-5980ab0ef3dc" (UID: "88c65caf-85af-4b62-bcb4-5980ab0ef3dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.123208 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.123263 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7z5ls\" (UniqueName: \"kubernetes.io/projected/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-kube-api-access-7z5ls\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.123286 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.123309 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88c65caf-85af-4b62-bcb4-5980ab0ef3dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.764941 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"88c65caf-85af-4b62-bcb4-5980ab0ef3dc","Type":"ContainerDied","Data":"fe5502aeb26e008e03231242e0f68c3678fe0d7d1cc0829d9a894c9d5902f2d5"} Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.765491 4857 scope.go:117] "RemoveContainer" containerID="c4c0303796540659ca714ede5624b707792775bc31659e86472eee7d7f67d243" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.765098 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.768435 4857 generic.go:334] "Generic (PLEG): container finished" podID="70f52857-7d57-421e-8bf0-b849a8c02ffe" containerID="f64fce28262cd259de4d3107ccf2fd6b7a1b3c2c8b4babaae8b518b5223d8aba" exitCode=0 Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.768475 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-m4rd7" event={"ID":"70f52857-7d57-421e-8bf0-b849a8c02ffe","Type":"ContainerDied","Data":"f64fce28262cd259de4d3107ccf2fd6b7a1b3c2c8b4babaae8b518b5223d8aba"} Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.805799 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.805845 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.821815 4857 scope.go:117] "RemoveContainer" containerID="e1eb7b2324f2b07c77d704df8935422c1179839daba017b69e5921a6b452affd" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.893006 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.893064 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.904560 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.919099 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:23 crc kubenswrapper[4857]: E1201 21:57:23.919594 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c65caf-85af-4b62-bcb4-5980ab0ef3dc" containerName="nova-metadata-log" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.919618 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c65caf-85af-4b62-bcb4-5980ab0ef3dc" containerName="nova-metadata-log" Dec 01 21:57:23 crc kubenswrapper[4857]: E1201 21:57:23.919651 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c65caf-85af-4b62-bcb4-5980ab0ef3dc" containerName="nova-metadata-metadata" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.919658 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c65caf-85af-4b62-bcb4-5980ab0ef3dc" containerName="nova-metadata-metadata" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.919834 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="88c65caf-85af-4b62-bcb4-5980ab0ef3dc" containerName="nova-metadata-metadata" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.919850 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="88c65caf-85af-4b62-bcb4-5980ab0ef3dc" containerName="nova-metadata-log" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.920961 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.923813 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.926357 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.934056 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.992861 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 21:57:23 crc kubenswrapper[4857]: I1201 21:57:23.992964 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.045904 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npznr\" (UniqueName: \"kubernetes.io/projected/1d8b05a0-713c-4d9a-a4d4-56431feac976-kube-api-access-npznr\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.046061 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d8b05a0-713c-4d9a-a4d4-56431feac976-logs\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.046100 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.046184 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.046391 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-config-data\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.148852 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npznr\" (UniqueName: \"kubernetes.io/projected/1d8b05a0-713c-4d9a-a4d4-56431feac976-kube-api-access-npznr\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.148974 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d8b05a0-713c-4d9a-a4d4-56431feac976-logs\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.149015 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.149157 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.149239 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-config-data\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.150336 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d8b05a0-713c-4d9a-a4d4-56431feac976-logs\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.159745 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-config-data\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.159804 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.167340 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.176952 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npznr\" (UniqueName: \"kubernetes.io/projected/1d8b05a0-713c-4d9a-a4d4-56431feac976-kube-api-access-npznr\") pod \"nova-metadata-0\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.241972 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.291233 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.343494 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.358607 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-jr4rg"] Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.358999 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" podUID="c76a0a21-8d25-4168-be04-941041bce44a" containerName="dnsmasq-dns" containerID="cri-o://557d60b969ae3ff9ad723c2be3a36ab7a1c599c3765e9e6b0b8aca470ed7cfee" gracePeriod=10 Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.824319 4857 generic.go:334] "Generic (PLEG): container finished" podID="c76a0a21-8d25-4168-be04-941041bce44a" containerID="557d60b969ae3ff9ad723c2be3a36ab7a1c599c3765e9e6b0b8aca470ed7cfee" exitCode=0 Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.824837 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" event={"ID":"c76a0a21-8d25-4168-be04-941041bce44a","Type":"ContainerDied","Data":"557d60b969ae3ff9ad723c2be3a36ab7a1c599c3765e9e6b0b8aca470ed7cfee"} Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.897580 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:24 crc kubenswrapper[4857]: W1201 21:57:24.923341 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d8b05a0_713c_4d9a_a4d4_56431feac976.slice/crio-cbf5f3dea8875dfd0570986f3d51069bef622801c099eea5d4d3d07437e951ec WatchSource:0}: Error finding container cbf5f3dea8875dfd0570986f3d51069bef622801c099eea5d4d3d07437e951ec: Status 404 returned error can't find the container with id cbf5f3dea8875dfd0570986f3d51069bef622801c099eea5d4d3d07437e951ec Dec 01 21:57:24 crc kubenswrapper[4857]: I1201 21:57:24.959564 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.084400 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c9e426f5-5fde-4067-8128-76f26f907fc8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.084396 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c9e426f5-5fde-4067-8128-76f26f907fc8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.102400 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.282052 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-dns-svc\") pod \"c76a0a21-8d25-4168-be04-941041bce44a\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.282495 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-config\") pod \"c76a0a21-8d25-4168-be04-941041bce44a\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.282575 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-dns-swift-storage-0\") pod \"c76a0a21-8d25-4168-be04-941041bce44a\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.282653 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdhkn\" (UniqueName: \"kubernetes.io/projected/c76a0a21-8d25-4168-be04-941041bce44a-kube-api-access-mdhkn\") pod \"c76a0a21-8d25-4168-be04-941041bce44a\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.282717 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-ovsdbserver-sb\") pod \"c76a0a21-8d25-4168-be04-941041bce44a\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.282771 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-ovsdbserver-nb\") pod \"c76a0a21-8d25-4168-be04-941041bce44a\" (UID: \"c76a0a21-8d25-4168-be04-941041bce44a\") " Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.320866 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c76a0a21-8d25-4168-be04-941041bce44a-kube-api-access-mdhkn" (OuterVolumeSpecName: "kube-api-access-mdhkn") pod "c76a0a21-8d25-4168-be04-941041bce44a" (UID: "c76a0a21-8d25-4168-be04-941041bce44a"). InnerVolumeSpecName "kube-api-access-mdhkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.385561 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdhkn\" (UniqueName: \"kubernetes.io/projected/c76a0a21-8d25-4168-be04-941041bce44a-kube-api-access-mdhkn\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.465572 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c76a0a21-8d25-4168-be04-941041bce44a" (UID: "c76a0a21-8d25-4168-be04-941041bce44a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.473223 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-config" (OuterVolumeSpecName: "config") pod "c76a0a21-8d25-4168-be04-941041bce44a" (UID: "c76a0a21-8d25-4168-be04-941041bce44a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.480947 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c76a0a21-8d25-4168-be04-941041bce44a" (UID: "c76a0a21-8d25-4168-be04-941041bce44a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.487780 4857 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.487810 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.487820 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.517277 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c76a0a21-8d25-4168-be04-941041bce44a" (UID: "c76a0a21-8d25-4168-be04-941041bce44a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.551094 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c76a0a21-8d25-4168-be04-941041bce44a" (UID: "c76a0a21-8d25-4168-be04-941041bce44a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.594419 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.594464 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c76a0a21-8d25-4168-be04-941041bce44a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.696968 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.801293 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-combined-ca-bundle\") pod \"70f52857-7d57-421e-8bf0-b849a8c02ffe\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.801966 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-config-data\") pod \"70f52857-7d57-421e-8bf0-b849a8c02ffe\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.802003 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-scripts\") pod \"70f52857-7d57-421e-8bf0-b849a8c02ffe\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.802104 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpl7q\" (UniqueName: \"kubernetes.io/projected/70f52857-7d57-421e-8bf0-b849a8c02ffe-kube-api-access-gpl7q\") pod \"70f52857-7d57-421e-8bf0-b849a8c02ffe\" (UID: \"70f52857-7d57-421e-8bf0-b849a8c02ffe\") " Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.807237 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70f52857-7d57-421e-8bf0-b849a8c02ffe-kube-api-access-gpl7q" (OuterVolumeSpecName: "kube-api-access-gpl7q") pod "70f52857-7d57-421e-8bf0-b849a8c02ffe" (UID: "70f52857-7d57-421e-8bf0-b849a8c02ffe"). InnerVolumeSpecName "kube-api-access-gpl7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.807980 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-scripts" (OuterVolumeSpecName: "scripts") pod "70f52857-7d57-421e-8bf0-b849a8c02ffe" (UID: "70f52857-7d57-421e-8bf0-b849a8c02ffe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.834917 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-config-data" (OuterVolumeSpecName: "config-data") pod "70f52857-7d57-421e-8bf0-b849a8c02ffe" (UID: "70f52857-7d57-421e-8bf0-b849a8c02ffe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.837329 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70f52857-7d57-421e-8bf0-b849a8c02ffe" (UID: "70f52857-7d57-421e-8bf0-b849a8c02ffe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.845904 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88c65caf-85af-4b62-bcb4-5980ab0ef3dc" path="/var/lib/kubelet/pods/88c65caf-85af-4b62-bcb4-5980ab0ef3dc/volumes" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.847021 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-m4rd7" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.849287 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.851141 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-m4rd7" event={"ID":"70f52857-7d57-421e-8bf0-b849a8c02ffe","Type":"ContainerDied","Data":"530f8de5bd12e9b9d3b0d768c71ab38f16d8997ec05bd4d1c26a35cd827a6e30"} Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.851197 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="530f8de5bd12e9b9d3b0d768c71ab38f16d8997ec05bd4d1c26a35cd827a6e30" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.851210 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-jr4rg" event={"ID":"c76a0a21-8d25-4168-be04-941041bce44a","Type":"ContainerDied","Data":"987f0ec2088516459cb591cc0aa07da579fd57acf4034c96946cfc5f88a75d7c"} Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.851232 4857 scope.go:117] "RemoveContainer" containerID="557d60b969ae3ff9ad723c2be3a36ab7a1c599c3765e9e6b0b8aca470ed7cfee" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.853561 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1d8b05a0-713c-4d9a-a4d4-56431feac976","Type":"ContainerStarted","Data":"a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71"} Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.853592 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1d8b05a0-713c-4d9a-a4d4-56431feac976","Type":"ContainerStarted","Data":"201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894"} Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.853601 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1d8b05a0-713c-4d9a-a4d4-56431feac976","Type":"ContainerStarted","Data":"cbf5f3dea8875dfd0570986f3d51069bef622801c099eea5d4d3d07437e951ec"} Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.888020 4857 scope.go:117] "RemoveContainer" containerID="7f814e0af9b78ad7d49afe09cdda869976239459e3508514ec5392487809be88" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.895000 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.894977459 podStartE2EDuration="2.894977459s" podCreationTimestamp="2025-12-01 21:57:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:57:25.877556757 +0000 UTC m=+1364.367619074" watchObservedRunningTime="2025-12-01 21:57:25.894977459 +0000 UTC m=+1364.385039776" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.906965 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpl7q\" (UniqueName: \"kubernetes.io/projected/70f52857-7d57-421e-8bf0-b849a8c02ffe-kube-api-access-gpl7q\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.906999 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.907011 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.907022 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f52857-7d57-421e-8bf0-b849a8c02ffe-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.948787 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-jr4rg"] Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.968253 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-jr4rg"] Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.976872 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.977949 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c9e426f5-5fde-4067-8128-76f26f907fc8" containerName="nova-api-log" containerID="cri-o://12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c" gracePeriod=30 Dec 01 21:57:25 crc kubenswrapper[4857]: I1201 21:57:25.978526 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c9e426f5-5fde-4067-8128-76f26f907fc8" containerName="nova-api-api" containerID="cri-o://d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752" gracePeriod=30 Dec 01 21:57:26 crc kubenswrapper[4857]: I1201 21:57:26.006866 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:26 crc kubenswrapper[4857]: I1201 21:57:26.091143 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:57:26 crc kubenswrapper[4857]: I1201 21:57:26.865683 4857 generic.go:334] "Generic (PLEG): container finished" podID="9df6644a-8870-44de-8e78-51f76e062f85" containerID="d78b0cfe9bd8d514b110c57c6cac5a8599248656c2b26a32beb8199cf1eae8c6" exitCode=0 Dec 01 21:57:26 crc kubenswrapper[4857]: I1201 21:57:26.865740 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-v9qs5" event={"ID":"9df6644a-8870-44de-8e78-51f76e062f85","Type":"ContainerDied","Data":"d78b0cfe9bd8d514b110c57c6cac5a8599248656c2b26a32beb8199cf1eae8c6"} Dec 01 21:57:26 crc kubenswrapper[4857]: I1201 21:57:26.868425 4857 generic.go:334] "Generic (PLEG): container finished" podID="c9e426f5-5fde-4067-8128-76f26f907fc8" containerID="12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c" exitCode=143 Dec 01 21:57:26 crc kubenswrapper[4857]: I1201 21:57:26.868512 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c9e426f5-5fde-4067-8128-76f26f907fc8","Type":"ContainerDied","Data":"12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c"} Dec 01 21:57:26 crc kubenswrapper[4857]: I1201 21:57:26.868639 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b" containerName="nova-scheduler-scheduler" containerID="cri-o://d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c" gracePeriod=30 Dec 01 21:57:27 crc kubenswrapper[4857]: I1201 21:57:27.850278 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c76a0a21-8d25-4168-be04-941041bce44a" path="/var/lib/kubelet/pods/c76a0a21-8d25-4168-be04-941041bce44a/volumes" Dec 01 21:57:27 crc kubenswrapper[4857]: I1201 21:57:27.880258 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1d8b05a0-713c-4d9a-a4d4-56431feac976" containerName="nova-metadata-log" containerID="cri-o://201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894" gracePeriod=30 Dec 01 21:57:27 crc kubenswrapper[4857]: I1201 21:57:27.880402 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1d8b05a0-713c-4d9a-a4d4-56431feac976" containerName="nova-metadata-metadata" containerID="cri-o://a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71" gracePeriod=30 Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.334725 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.455352 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-scripts\") pod \"9df6644a-8870-44de-8e78-51f76e062f85\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.455469 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-config-data\") pod \"9df6644a-8870-44de-8e78-51f76e062f85\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.455497 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zffkc\" (UniqueName: \"kubernetes.io/projected/9df6644a-8870-44de-8e78-51f76e062f85-kube-api-access-zffkc\") pod \"9df6644a-8870-44de-8e78-51f76e062f85\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.455606 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-combined-ca-bundle\") pod \"9df6644a-8870-44de-8e78-51f76e062f85\" (UID: \"9df6644a-8870-44de-8e78-51f76e062f85\") " Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.463791 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9df6644a-8870-44de-8e78-51f76e062f85-kube-api-access-zffkc" (OuterVolumeSpecName: "kube-api-access-zffkc") pod "9df6644a-8870-44de-8e78-51f76e062f85" (UID: "9df6644a-8870-44de-8e78-51f76e062f85"). InnerVolumeSpecName "kube-api-access-zffkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.464862 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-scripts" (OuterVolumeSpecName: "scripts") pod "9df6644a-8870-44de-8e78-51f76e062f85" (UID: "9df6644a-8870-44de-8e78-51f76e062f85"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.484266 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.491110 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-config-data" (OuterVolumeSpecName: "config-data") pod "9df6644a-8870-44de-8e78-51f76e062f85" (UID: "9df6644a-8870-44de-8e78-51f76e062f85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.496322 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9df6644a-8870-44de-8e78-51f76e062f85" (UID: "9df6644a-8870-44de-8e78-51f76e062f85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.558478 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zffkc\" (UniqueName: \"kubernetes.io/projected/9df6644a-8870-44de-8e78-51f76e062f85-kube-api-access-zffkc\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.558519 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.558539 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.558550 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9df6644a-8870-44de-8e78-51f76e062f85-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.660273 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-combined-ca-bundle\") pod \"1d8b05a0-713c-4d9a-a4d4-56431feac976\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.660366 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-nova-metadata-tls-certs\") pod \"1d8b05a0-713c-4d9a-a4d4-56431feac976\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.660504 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npznr\" (UniqueName: \"kubernetes.io/projected/1d8b05a0-713c-4d9a-a4d4-56431feac976-kube-api-access-npznr\") pod \"1d8b05a0-713c-4d9a-a4d4-56431feac976\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.660569 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-config-data\") pod \"1d8b05a0-713c-4d9a-a4d4-56431feac976\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.660665 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d8b05a0-713c-4d9a-a4d4-56431feac976-logs\") pod \"1d8b05a0-713c-4d9a-a4d4-56431feac976\" (UID: \"1d8b05a0-713c-4d9a-a4d4-56431feac976\") " Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.661568 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d8b05a0-713c-4d9a-a4d4-56431feac976-logs" (OuterVolumeSpecName: "logs") pod "1d8b05a0-713c-4d9a-a4d4-56431feac976" (UID: "1d8b05a0-713c-4d9a-a4d4-56431feac976"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.666050 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d8b05a0-713c-4d9a-a4d4-56431feac976-kube-api-access-npznr" (OuterVolumeSpecName: "kube-api-access-npznr") pod "1d8b05a0-713c-4d9a-a4d4-56431feac976" (UID: "1d8b05a0-713c-4d9a-a4d4-56431feac976"). InnerVolumeSpecName "kube-api-access-npznr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.699810 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d8b05a0-713c-4d9a-a4d4-56431feac976" (UID: "1d8b05a0-713c-4d9a-a4d4-56431feac976"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.710540 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-config-data" (OuterVolumeSpecName: "config-data") pod "1d8b05a0-713c-4d9a-a4d4-56431feac976" (UID: "1d8b05a0-713c-4d9a-a4d4-56431feac976"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.745282 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1d8b05a0-713c-4d9a-a4d4-56431feac976" (UID: "1d8b05a0-713c-4d9a-a4d4-56431feac976"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.763182 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d8b05a0-713c-4d9a-a4d4-56431feac976-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.763229 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.763246 4857 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.763263 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npznr\" (UniqueName: \"kubernetes.io/projected/1d8b05a0-713c-4d9a-a4d4-56431feac976-kube-api-access-npznr\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.763276 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d8b05a0-713c-4d9a-a4d4-56431feac976-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:28 crc kubenswrapper[4857]: E1201 21:57:28.808247 4857 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 21:57:28 crc kubenswrapper[4857]: E1201 21:57:28.810817 4857 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 21:57:28 crc kubenswrapper[4857]: E1201 21:57:28.812950 4857 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 21:57:28 crc kubenswrapper[4857]: E1201 21:57:28.813023 4857 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b" containerName="nova-scheduler-scheduler" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.893766 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-v9qs5" event={"ID":"9df6644a-8870-44de-8e78-51f76e062f85","Type":"ContainerDied","Data":"739fc919a4043e77d470a4bc525f5f5bc673415c029400f17fdcf3a09aaff914"} Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.893819 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="739fc919a4043e77d470a4bc525f5f5bc673415c029400f17fdcf3a09aaff914" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.893893 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-v9qs5" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.899497 4857 generic.go:334] "Generic (PLEG): container finished" podID="1d8b05a0-713c-4d9a-a4d4-56431feac976" containerID="a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71" exitCode=0 Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.899537 4857 generic.go:334] "Generic (PLEG): container finished" podID="1d8b05a0-713c-4d9a-a4d4-56431feac976" containerID="201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894" exitCode=143 Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.899561 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1d8b05a0-713c-4d9a-a4d4-56431feac976","Type":"ContainerDied","Data":"a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71"} Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.899595 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1d8b05a0-713c-4d9a-a4d4-56431feac976","Type":"ContainerDied","Data":"201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894"} Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.899609 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1d8b05a0-713c-4d9a-a4d4-56431feac976","Type":"ContainerDied","Data":"cbf5f3dea8875dfd0570986f3d51069bef622801c099eea5d4d3d07437e951ec"} Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.899628 4857 scope.go:117] "RemoveContainer" containerID="a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.899796 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:57:28 crc kubenswrapper[4857]: I1201 21:57:28.963009 4857 scope.go:117] "RemoveContainer" containerID="201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.009180 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 21:57:29 crc kubenswrapper[4857]: E1201 21:57:29.009702 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8b05a0-713c-4d9a-a4d4-56431feac976" containerName="nova-metadata-metadata" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.009723 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8b05a0-713c-4d9a-a4d4-56431feac976" containerName="nova-metadata-metadata" Dec 01 21:57:29 crc kubenswrapper[4857]: E1201 21:57:29.009746 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c76a0a21-8d25-4168-be04-941041bce44a" containerName="dnsmasq-dns" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.009754 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c76a0a21-8d25-4168-be04-941041bce44a" containerName="dnsmasq-dns" Dec 01 21:57:29 crc kubenswrapper[4857]: E1201 21:57:29.009767 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8b05a0-713c-4d9a-a4d4-56431feac976" containerName="nova-metadata-log" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.009774 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8b05a0-713c-4d9a-a4d4-56431feac976" containerName="nova-metadata-log" Dec 01 21:57:29 crc kubenswrapper[4857]: E1201 21:57:29.009792 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c76a0a21-8d25-4168-be04-941041bce44a" containerName="init" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.009799 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c76a0a21-8d25-4168-be04-941041bce44a" containerName="init" Dec 01 21:57:29 crc kubenswrapper[4857]: E1201 21:57:29.009820 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df6644a-8870-44de-8e78-51f76e062f85" containerName="nova-cell1-conductor-db-sync" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.009828 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df6644a-8870-44de-8e78-51f76e062f85" containerName="nova-cell1-conductor-db-sync" Dec 01 21:57:29 crc kubenswrapper[4857]: E1201 21:57:29.009852 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f52857-7d57-421e-8bf0-b849a8c02ffe" containerName="nova-manage" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.009860 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f52857-7d57-421e-8bf0-b849a8c02ffe" containerName="nova-manage" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.010107 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d8b05a0-713c-4d9a-a4d4-56431feac976" containerName="nova-metadata-log" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.010136 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d8b05a0-713c-4d9a-a4d4-56431feac976" containerName="nova-metadata-metadata" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.010149 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c76a0a21-8d25-4168-be04-941041bce44a" containerName="dnsmasq-dns" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.010168 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9df6644a-8870-44de-8e78-51f76e062f85" containerName="nova-cell1-conductor-db-sync" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.010182 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f52857-7d57-421e-8bf0-b849a8c02ffe" containerName="nova-manage" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.013596 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.025404 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.026815 4857 scope.go:117] "RemoveContainer" containerID="a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.026944 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 21:57:29 crc kubenswrapper[4857]: E1201 21:57:29.027313 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71\": container with ID starting with a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71 not found: ID does not exist" containerID="a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.027347 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71"} err="failed to get container status \"a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71\": rpc error: code = NotFound desc = could not find container \"a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71\": container with ID starting with a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71 not found: ID does not exist" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.027370 4857 scope.go:117] "RemoveContainer" containerID="201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894" Dec 01 21:57:29 crc kubenswrapper[4857]: E1201 21:57:29.027658 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894\": container with ID starting with 201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894 not found: ID does not exist" containerID="201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.027708 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894"} err="failed to get container status \"201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894\": rpc error: code = NotFound desc = could not find container \"201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894\": container with ID starting with 201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894 not found: ID does not exist" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.027744 4857 scope.go:117] "RemoveContainer" containerID="a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.028100 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71"} err="failed to get container status \"a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71\": rpc error: code = NotFound desc = could not find container \"a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71\": container with ID starting with a2d43b8af2d11910cb35ff457df7e12594879e6b188cd208628f2f02e4e31e71 not found: ID does not exist" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.028125 4857 scope.go:117] "RemoveContainer" containerID="201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.028367 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894"} err="failed to get container status \"201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894\": rpc error: code = NotFound desc = could not find container \"201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894\": container with ID starting with 201a3a013fe8f1fbc1f6dd4f5e43311eeca3ae79faa75e3587cd434a26bf3894 not found: ID does not exist" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.035898 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.044966 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.052571 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.055149 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.057587 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.058561 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.062022 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.172181 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.172223 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0a02ec-0eec-4459-9de2-96f044ff45c7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cd0a02ec-0eec-4459-9de2-96f044ff45c7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.172255 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzqh2\" (UniqueName: \"kubernetes.io/projected/14d98749-4c8f-4ad4-abc4-3229198de8f7-kube-api-access-nzqh2\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.172296 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-config-data\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.172311 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrmgc\" (UniqueName: \"kubernetes.io/projected/cd0a02ec-0eec-4459-9de2-96f044ff45c7-kube-api-access-zrmgc\") pod \"nova-cell1-conductor-0\" (UID: \"cd0a02ec-0eec-4459-9de2-96f044ff45c7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.172400 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0a02ec-0eec-4459-9de2-96f044ff45c7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cd0a02ec-0eec-4459-9de2-96f044ff45c7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.172451 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.172526 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14d98749-4c8f-4ad4-abc4-3229198de8f7-logs\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.274513 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0a02ec-0eec-4459-9de2-96f044ff45c7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cd0a02ec-0eec-4459-9de2-96f044ff45c7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.274578 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzqh2\" (UniqueName: \"kubernetes.io/projected/14d98749-4c8f-4ad4-abc4-3229198de8f7-kube-api-access-nzqh2\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.274614 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-config-data\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.274669 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrmgc\" (UniqueName: \"kubernetes.io/projected/cd0a02ec-0eec-4459-9de2-96f044ff45c7-kube-api-access-zrmgc\") pod \"nova-cell1-conductor-0\" (UID: \"cd0a02ec-0eec-4459-9de2-96f044ff45c7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.274735 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0a02ec-0eec-4459-9de2-96f044ff45c7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cd0a02ec-0eec-4459-9de2-96f044ff45c7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.274772 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.274877 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14d98749-4c8f-4ad4-abc4-3229198de8f7-logs\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.274964 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.275588 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14d98749-4c8f-4ad4-abc4-3229198de8f7-logs\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.279729 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-config-data\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.279765 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.279868 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0a02ec-0eec-4459-9de2-96f044ff45c7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cd0a02ec-0eec-4459-9de2-96f044ff45c7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.292177 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.296679 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzqh2\" (UniqueName: \"kubernetes.io/projected/14d98749-4c8f-4ad4-abc4-3229198de8f7-kube-api-access-nzqh2\") pod \"nova-metadata-0\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.296741 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0a02ec-0eec-4459-9de2-96f044ff45c7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cd0a02ec-0eec-4459-9de2-96f044ff45c7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.305578 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrmgc\" (UniqueName: \"kubernetes.io/projected/cd0a02ec-0eec-4459-9de2-96f044ff45c7-kube-api-access-zrmgc\") pod \"nova-cell1-conductor-0\" (UID: \"cd0a02ec-0eec-4459-9de2-96f044ff45c7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.377665 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.386418 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.474830 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.583522 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-config-data\") pod \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\" (UID: \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\") " Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.583607 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svmlg\" (UniqueName: \"kubernetes.io/projected/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-kube-api-access-svmlg\") pod \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\" (UID: \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\") " Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.583672 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-combined-ca-bundle\") pod \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\" (UID: \"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b\") " Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.590620 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-kube-api-access-svmlg" (OuterVolumeSpecName: "kube-api-access-svmlg") pod "2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b" (UID: "2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b"). InnerVolumeSpecName "kube-api-access-svmlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.617553 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b" (UID: "2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.619826 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-config-data" (OuterVolumeSpecName: "config-data") pod "2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b" (UID: "2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.686059 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.686106 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svmlg\" (UniqueName: \"kubernetes.io/projected/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-kube-api-access-svmlg\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.686126 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.850660 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d8b05a0-713c-4d9a-a4d4-56431feac976" path="/var/lib/kubelet/pods/1d8b05a0-713c-4d9a-a4d4-56431feac976/volumes" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.894345 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.911796 4857 generic.go:334] "Generic (PLEG): container finished" podID="2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b" containerID="d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c" exitCode=0 Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.911870 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b","Type":"ContainerDied","Data":"d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c"} Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.911882 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.911902 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b","Type":"ContainerDied","Data":"7a9cdce76094fedcccb39a4ffc860de9c3f53a398bfca8c494393844b65d8a3b"} Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.911919 4857 scope.go:117] "RemoveContainer" containerID="d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.912922 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cd0a02ec-0eec-4459-9de2-96f044ff45c7","Type":"ContainerStarted","Data":"cb27717ec0b7e5f5b7430be9d6ea047de3d2acb398d763df7add29b84fcc2884"} Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.939568 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.951611 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.965183 4857 scope.go:117] "RemoveContainer" containerID="d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c" Dec 01 21:57:29 crc kubenswrapper[4857]: E1201 21:57:29.967968 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c\": container with ID starting with d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c not found: ID does not exist" containerID="d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.968013 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c"} err="failed to get container status \"d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c\": rpc error: code = NotFound desc = could not find container \"d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c\": container with ID starting with d53462b443da4993db22d43eb8bc6c87470ff3176e270d0d4c2d4b6e07c7c50c not found: ID does not exist" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.971785 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:57:29 crc kubenswrapper[4857]: E1201 21:57:29.972473 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b" containerName="nova-scheduler-scheduler" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.972497 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b" containerName="nova-scheduler-scheduler" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.972729 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b" containerName="nova-scheduler-scheduler" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.974126 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.977201 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 21:57:29 crc kubenswrapper[4857]: I1201 21:57:29.985538 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.010263 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:57:30 crc kubenswrapper[4857]: W1201 21:57:30.025601 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14d98749_4c8f_4ad4_abc4_3229198de8f7.slice/crio-6d9510d9fd0b63ed185c4b9eb0bcd786069d2f47b2d50353e845e8a89d36b1d3 WatchSource:0}: Error finding container 6d9510d9fd0b63ed185c4b9eb0bcd786069d2f47b2d50353e845e8a89d36b1d3: Status 404 returned error can't find the container with id 6d9510d9fd0b63ed185c4b9eb0bcd786069d2f47b2d50353e845e8a89d36b1d3 Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.097418 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwpx4\" (UniqueName: \"kubernetes.io/projected/c402ef80-d243-4a46-bd8f-9fa0036d771e-kube-api-access-kwpx4\") pod \"nova-scheduler-0\" (UID: \"c402ef80-d243-4a46-bd8f-9fa0036d771e\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.097711 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c402ef80-d243-4a46-bd8f-9fa0036d771e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c402ef80-d243-4a46-bd8f-9fa0036d771e\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.097932 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c402ef80-d243-4a46-bd8f-9fa0036d771e-config-data\") pod \"nova-scheduler-0\" (UID: \"c402ef80-d243-4a46-bd8f-9fa0036d771e\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.200796 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c402ef80-d243-4a46-bd8f-9fa0036d771e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c402ef80-d243-4a46-bd8f-9fa0036d771e\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.202288 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c402ef80-d243-4a46-bd8f-9fa0036d771e-config-data\") pod \"nova-scheduler-0\" (UID: \"c402ef80-d243-4a46-bd8f-9fa0036d771e\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.202729 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwpx4\" (UniqueName: \"kubernetes.io/projected/c402ef80-d243-4a46-bd8f-9fa0036d771e-kube-api-access-kwpx4\") pod \"nova-scheduler-0\" (UID: \"c402ef80-d243-4a46-bd8f-9fa0036d771e\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.207960 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c402ef80-d243-4a46-bd8f-9fa0036d771e-config-data\") pod \"nova-scheduler-0\" (UID: \"c402ef80-d243-4a46-bd8f-9fa0036d771e\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.218604 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c402ef80-d243-4a46-bd8f-9fa0036d771e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c402ef80-d243-4a46-bd8f-9fa0036d771e\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.225217 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwpx4\" (UniqueName: \"kubernetes.io/projected/c402ef80-d243-4a46-bd8f-9fa0036d771e-kube-api-access-kwpx4\") pod \"nova-scheduler-0\" (UID: \"c402ef80-d243-4a46-bd8f-9fa0036d771e\") " pod="openstack/nova-scheduler-0" Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.479283 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.909608 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.927881 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"14d98749-4c8f-4ad4-abc4-3229198de8f7","Type":"ContainerStarted","Data":"911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f"} Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.927945 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"14d98749-4c8f-4ad4-abc4-3229198de8f7","Type":"ContainerStarted","Data":"8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664"} Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.927958 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"14d98749-4c8f-4ad4-abc4-3229198de8f7","Type":"ContainerStarted","Data":"6d9510d9fd0b63ed185c4b9eb0bcd786069d2f47b2d50353e845e8a89d36b1d3"} Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.956973 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cd0a02ec-0eec-4459-9de2-96f044ff45c7","Type":"ContainerStarted","Data":"6d1eeac14c64f834dd6b03452c5726e542086e3aa4569e50487714df3e23d900"} Dec 01 21:57:30 crc kubenswrapper[4857]: I1201 21:57:30.958985 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.020417 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e426f5-5fde-4067-8128-76f26f907fc8-combined-ca-bundle\") pod \"c9e426f5-5fde-4067-8128-76f26f907fc8\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.020469 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e426f5-5fde-4067-8128-76f26f907fc8-config-data\") pod \"c9e426f5-5fde-4067-8128-76f26f907fc8\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.020532 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9e426f5-5fde-4067-8128-76f26f907fc8-logs\") pod \"c9e426f5-5fde-4067-8128-76f26f907fc8\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.020923 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l4vh\" (UniqueName: \"kubernetes.io/projected/c9e426f5-5fde-4067-8128-76f26f907fc8-kube-api-access-4l4vh\") pod \"c9e426f5-5fde-4067-8128-76f26f907fc8\" (UID: \"c9e426f5-5fde-4067-8128-76f26f907fc8\") " Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.057011 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9e426f5-5fde-4067-8128-76f26f907fc8-logs" (OuterVolumeSpecName: "logs") pod "c9e426f5-5fde-4067-8128-76f26f907fc8" (UID: "c9e426f5-5fde-4067-8128-76f26f907fc8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.070691 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9e426f5-5fde-4067-8128-76f26f907fc8-kube-api-access-4l4vh" (OuterVolumeSpecName: "kube-api-access-4l4vh") pod "c9e426f5-5fde-4067-8128-76f26f907fc8" (UID: "c9e426f5-5fde-4067-8128-76f26f907fc8"). InnerVolumeSpecName "kube-api-access-4l4vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.100500 4857 generic.go:334] "Generic (PLEG): container finished" podID="c9e426f5-5fde-4067-8128-76f26f907fc8" containerID="d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752" exitCode=0 Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.100553 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c9e426f5-5fde-4067-8128-76f26f907fc8","Type":"ContainerDied","Data":"d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752"} Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.100582 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c9e426f5-5fde-4067-8128-76f26f907fc8","Type":"ContainerDied","Data":"dc1ca86002a0f1e47ab5646e6827a820f0eaa87a31454c2253efda60cb2f29bf"} Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.100617 4857 scope.go:117] "RemoveContainer" containerID="d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.100873 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.113864 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9e426f5-5fde-4067-8128-76f26f907fc8-config-data" (OuterVolumeSpecName: "config-data") pod "c9e426f5-5fde-4067-8128-76f26f907fc8" (UID: "c9e426f5-5fde-4067-8128-76f26f907fc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.127784 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l4vh\" (UniqueName: \"kubernetes.io/projected/c9e426f5-5fde-4067-8128-76f26f907fc8-kube-api-access-4l4vh\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.128456 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e426f5-5fde-4067-8128-76f26f907fc8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.128495 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9e426f5-5fde-4067-8128-76f26f907fc8-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.137950 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9e426f5-5fde-4067-8128-76f26f907fc8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9e426f5-5fde-4067-8128-76f26f907fc8" (UID: "c9e426f5-5fde-4067-8128-76f26f907fc8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.143179 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.143156269 podStartE2EDuration="3.143156269s" podCreationTimestamp="2025-12-01 21:57:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:57:31.080433245 +0000 UTC m=+1369.570495562" watchObservedRunningTime="2025-12-01 21:57:31.143156269 +0000 UTC m=+1369.633218576" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.147485 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.154647 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.154635463 podStartE2EDuration="3.154635463s" podCreationTimestamp="2025-12-01 21:57:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:57:31.099790834 +0000 UTC m=+1369.589853151" watchObservedRunningTime="2025-12-01 21:57:31.154635463 +0000 UTC m=+1369.644697780" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.230997 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e426f5-5fde-4067-8128-76f26f907fc8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.312091 4857 scope.go:117] "RemoveContainer" containerID="12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.346424 4857 scope.go:117] "RemoveContainer" containerID="d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752" Dec 01 21:57:31 crc kubenswrapper[4857]: E1201 21:57:31.347745 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752\": container with ID starting with d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752 not found: ID does not exist" containerID="d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.347810 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752"} err="failed to get container status \"d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752\": rpc error: code = NotFound desc = could not find container \"d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752\": container with ID starting with d74f0bee0777fdb6267f18c571754447d77138076b1d08c36edc994c43350752 not found: ID does not exist" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.347847 4857 scope.go:117] "RemoveContainer" containerID="12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c" Dec 01 21:57:31 crc kubenswrapper[4857]: E1201 21:57:31.348269 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c\": container with ID starting with 12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c not found: ID does not exist" containerID="12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.348293 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c"} err="failed to get container status \"12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c\": rpc error: code = NotFound desc = could not find container \"12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c\": container with ID starting with 12cdd4eeefb1cf74e73466fea15542983d14155567e88d5eadc14192b0f7457c not found: ID does not exist" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.445517 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.465822 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.479158 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:31 crc kubenswrapper[4857]: E1201 21:57:31.479900 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e426f5-5fde-4067-8128-76f26f907fc8" containerName="nova-api-api" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.479928 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e426f5-5fde-4067-8128-76f26f907fc8" containerName="nova-api-api" Dec 01 21:57:31 crc kubenswrapper[4857]: E1201 21:57:31.479951 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e426f5-5fde-4067-8128-76f26f907fc8" containerName="nova-api-log" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.479979 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e426f5-5fde-4067-8128-76f26f907fc8" containerName="nova-api-log" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.480231 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9e426f5-5fde-4067-8128-76f26f907fc8" containerName="nova-api-log" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.480278 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9e426f5-5fde-4067-8128-76f26f907fc8" containerName="nova-api-api" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.481648 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.491729 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.497646 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.538031 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.538188 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-config-data\") pod \"nova-api-0\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.538233 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-logs\") pod \"nova-api-0\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.538285 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4wfb\" (UniqueName: \"kubernetes.io/projected/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-kube-api-access-j4wfb\") pod \"nova-api-0\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.641923 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-config-data\") pod \"nova-api-0\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.642020 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-logs\") pod \"nova-api-0\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.642107 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4wfb\" (UniqueName: \"kubernetes.io/projected/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-kube-api-access-j4wfb\") pod \"nova-api-0\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.642449 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.643344 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-logs\") pod \"nova-api-0\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.648304 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-config-data\") pod \"nova-api-0\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.664386 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.673814 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4wfb\" (UniqueName: \"kubernetes.io/projected/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-kube-api-access-j4wfb\") pod \"nova-api-0\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.823280 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.857541 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b" path="/var/lib/kubelet/pods/2e2dee15-4cb9-49cc-9e7e-1ff88c0e0c4b/volumes" Dec 01 21:57:31 crc kubenswrapper[4857]: I1201 21:57:31.859778 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9e426f5-5fde-4067-8128-76f26f907fc8" path="/var/lib/kubelet/pods/c9e426f5-5fde-4067-8128-76f26f907fc8/volumes" Dec 01 21:57:32 crc kubenswrapper[4857]: I1201 21:57:32.113195 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c402ef80-d243-4a46-bd8f-9fa0036d771e","Type":"ContainerStarted","Data":"3931aa3a7a90531868722bb8b84722451b0a11cc16dd4fdc24674887903a3cd3"} Dec 01 21:57:32 crc kubenswrapper[4857]: I1201 21:57:32.113606 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c402ef80-d243-4a46-bd8f-9fa0036d771e","Type":"ContainerStarted","Data":"3d1667229fe8bee04af3b8ac172f200d61a20f5d788e37a02e49f6be922d3a05"} Dec 01 21:57:32 crc kubenswrapper[4857]: I1201 21:57:32.133423 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.133379341 podStartE2EDuration="3.133379341s" podCreationTimestamp="2025-12-01 21:57:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:57:32.1285066 +0000 UTC m=+1370.618568917" watchObservedRunningTime="2025-12-01 21:57:32.133379341 +0000 UTC m=+1370.623441658" Dec 01 21:57:32 crc kubenswrapper[4857]: W1201 21:57:32.299081 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb9754b2_56a5_4a72_a3fa_3d6c280027f4.slice/crio-92ba9cd6d5a36a000bef2f2922ac9c43fc3ce5609c1f179dbdef059c5f0bfd06 WatchSource:0}: Error finding container 92ba9cd6d5a36a000bef2f2922ac9c43fc3ce5609c1f179dbdef059c5f0bfd06: Status 404 returned error can't find the container with id 92ba9cd6d5a36a000bef2f2922ac9c43fc3ce5609c1f179dbdef059c5f0bfd06 Dec 01 21:57:32 crc kubenswrapper[4857]: I1201 21:57:32.300660 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:33 crc kubenswrapper[4857]: I1201 21:57:33.128348 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eb9754b2-56a5-4a72-a3fa-3d6c280027f4","Type":"ContainerStarted","Data":"5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c"} Dec 01 21:57:33 crc kubenswrapper[4857]: I1201 21:57:33.128809 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eb9754b2-56a5-4a72-a3fa-3d6c280027f4","Type":"ContainerStarted","Data":"21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be"} Dec 01 21:57:33 crc kubenswrapper[4857]: I1201 21:57:33.128834 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eb9754b2-56a5-4a72-a3fa-3d6c280027f4","Type":"ContainerStarted","Data":"92ba9cd6d5a36a000bef2f2922ac9c43fc3ce5609c1f179dbdef059c5f0bfd06"} Dec 01 21:57:33 crc kubenswrapper[4857]: I1201 21:57:33.173313 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.173290064 podStartE2EDuration="2.173290064s" podCreationTimestamp="2025-12-01 21:57:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:57:33.158871817 +0000 UTC m=+1371.648934134" watchObservedRunningTime="2025-12-01 21:57:33.173290064 +0000 UTC m=+1371.663352381" Dec 01 21:57:34 crc kubenswrapper[4857]: I1201 21:57:34.385530 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 21:57:34 crc kubenswrapper[4857]: I1201 21:57:34.387301 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 21:57:35 crc kubenswrapper[4857]: I1201 21:57:35.479741 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 21:57:39 crc kubenswrapper[4857]: I1201 21:57:39.385464 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 21:57:39 crc kubenswrapper[4857]: I1201 21:57:39.386079 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 21:57:39 crc kubenswrapper[4857]: I1201 21:57:39.412582 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 01 21:57:40 crc kubenswrapper[4857]: I1201 21:57:40.404337 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 21:57:40 crc kubenswrapper[4857]: I1201 21:57:40.404378 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 21:57:40 crc kubenswrapper[4857]: I1201 21:57:40.480563 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 21:57:40 crc kubenswrapper[4857]: I1201 21:57:40.539430 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 21:57:41 crc kubenswrapper[4857]: I1201 21:57:41.260791 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 21:57:41 crc kubenswrapper[4857]: I1201 21:57:41.824067 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 21:57:41 crc kubenswrapper[4857]: I1201 21:57:41.824148 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 21:57:42 crc kubenswrapper[4857]: I1201 21:57:42.911396 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 21:57:42 crc kubenswrapper[4857]: I1201 21:57:42.911603 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 21:57:43 crc kubenswrapper[4857]: I1201 21:57:43.874545 4857 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod070fb164-a99c-481e-8e3d-4dbae13e2073"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod070fb164-a99c-481e-8e3d-4dbae13e2073] : Timed out while waiting for systemd to remove kubepods-besteffort-pod070fb164_a99c_481e_8e3d_4dbae13e2073.slice" Dec 01 21:57:43 crc kubenswrapper[4857]: E1201 21:57:43.874607 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod070fb164-a99c-481e-8e3d-4dbae13e2073] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod070fb164-a99c-481e-8e3d-4dbae13e2073] : Timed out while waiting for systemd to remove kubepods-besteffort-pod070fb164_a99c_481e_8e3d_4dbae13e2073.slice" pod="openstack/ceilometer-0" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.250494 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.294511 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.315660 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.329928 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.341779 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.347202 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.363153 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.365308 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.409467 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.409591 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-config-data\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.409627 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgxxr\" (UniqueName: \"kubernetes.io/projected/bad773e6-67bc-4260-bb1a-f56ab753c7fc-kube-api-access-pgxxr\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.409704 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bad773e6-67bc-4260-bb1a-f56ab753c7fc-run-httpd\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.409837 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.409861 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-scripts\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.409913 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bad773e6-67bc-4260-bb1a-f56ab753c7fc-log-httpd\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.511654 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-config-data\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.511704 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgxxr\" (UniqueName: \"kubernetes.io/projected/bad773e6-67bc-4260-bb1a-f56ab753c7fc-kube-api-access-pgxxr\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.511723 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bad773e6-67bc-4260-bb1a-f56ab753c7fc-run-httpd\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.511812 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.511877 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-scripts\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.511914 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bad773e6-67bc-4260-bb1a-f56ab753c7fc-log-httpd\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.511958 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.512825 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bad773e6-67bc-4260-bb1a-f56ab753c7fc-run-httpd\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.513570 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bad773e6-67bc-4260-bb1a-f56ab753c7fc-log-httpd\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.520140 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.521302 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-config-data\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.521560 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-scripts\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.532433 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.533066 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgxxr\" (UniqueName: \"kubernetes.io/projected/bad773e6-67bc-4260-bb1a-f56ab753c7fc-kube-api-access-pgxxr\") pod \"ceilometer-0\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " pod="openstack/ceilometer-0" Dec 01 21:57:44 crc kubenswrapper[4857]: I1201 21:57:44.679033 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:57:45 crc kubenswrapper[4857]: I1201 21:57:45.319465 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:57:45 crc kubenswrapper[4857]: I1201 21:57:45.851903 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="070fb164-a99c-481e-8e3d-4dbae13e2073" path="/var/lib/kubelet/pods/070fb164-a99c-481e-8e3d-4dbae13e2073/volumes" Dec 01 21:57:46 crc kubenswrapper[4857]: I1201 21:57:46.290342 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bad773e6-67bc-4260-bb1a-f56ab753c7fc","Type":"ContainerStarted","Data":"141bc249b23cc089bad1b37bf8cfb2dbe70492872417fc7f44607080c7a6326c"} Dec 01 21:57:46 crc kubenswrapper[4857]: I1201 21:57:46.291273 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bad773e6-67bc-4260-bb1a-f56ab753c7fc","Type":"ContainerStarted","Data":"7ab4aac76c05dda4a1aa7c9371054ddedce2f75d64556d96d3f0fc2039182abf"} Dec 01 21:57:47 crc kubenswrapper[4857]: I1201 21:57:47.314961 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bad773e6-67bc-4260-bb1a-f56ab753c7fc","Type":"ContainerStarted","Data":"cb1d06839a70f4d244af67fafd625326bc4e53c6cc8e693fad096bb8ae347a32"} Dec 01 21:57:48 crc kubenswrapper[4857]: I1201 21:57:48.345059 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bad773e6-67bc-4260-bb1a-f56ab753c7fc","Type":"ContainerStarted","Data":"88e00e4e163d08859215bbed9e0f01ff178768407accda52fd4e176b8362fe31"} Dec 01 21:57:49 crc kubenswrapper[4857]: I1201 21:57:49.356787 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bad773e6-67bc-4260-bb1a-f56ab753c7fc","Type":"ContainerStarted","Data":"888eed7f0dfc676d49e1922a13d2154d16c323976192546fb6ec67eb083a9dc7"} Dec 01 21:57:49 crc kubenswrapper[4857]: I1201 21:57:49.357338 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 21:57:49 crc kubenswrapper[4857]: I1201 21:57:49.394675 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9732276629999999 podStartE2EDuration="5.394650938s" podCreationTimestamp="2025-12-01 21:57:44 +0000 UTC" firstStartedPulling="2025-12-01 21:57:45.324262977 +0000 UTC m=+1383.814325294" lastFinishedPulling="2025-12-01 21:57:48.745686212 +0000 UTC m=+1387.235748569" observedRunningTime="2025-12-01 21:57:49.383364079 +0000 UTC m=+1387.873426396" watchObservedRunningTime="2025-12-01 21:57:49.394650938 +0000 UTC m=+1387.884713285" Dec 01 21:57:49 crc kubenswrapper[4857]: I1201 21:57:49.400406 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 21:57:49 crc kubenswrapper[4857]: I1201 21:57:49.401717 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 21:57:49 crc kubenswrapper[4857]: I1201 21:57:49.412568 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 21:57:50 crc kubenswrapper[4857]: I1201 21:57:50.384581 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 21:57:51 crc kubenswrapper[4857]: I1201 21:57:51.830075 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 21:57:51 crc kubenswrapper[4857]: I1201 21:57:51.832114 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 21:57:51 crc kubenswrapper[4857]: I1201 21:57:51.834235 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 21:57:51 crc kubenswrapper[4857]: I1201 21:57:51.906178 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.345975 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.399262 4857 generic.go:334] "Generic (PLEG): container finished" podID="562073c1-2475-4e08-a121-b69cf751fa63" containerID="202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce" exitCode=137 Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.399672 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"562073c1-2475-4e08-a121-b69cf751fa63","Type":"ContainerDied","Data":"202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce"} Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.399858 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"562073c1-2475-4e08-a121-b69cf751fa63","Type":"ContainerDied","Data":"e87be741b0a475c9cf4b961d4cd0938e07dfa19c74eedfcc73dc2da97429867d"} Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.400019 4857 scope.go:117] "RemoveContainer" containerID="202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.400607 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.400929 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.404624 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.438636 4857 scope.go:117] "RemoveContainer" containerID="202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce" Dec 01 21:57:52 crc kubenswrapper[4857]: E1201 21:57:52.439284 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce\": container with ID starting with 202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce not found: ID does not exist" containerID="202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.439338 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce"} err="failed to get container status \"202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce\": rpc error: code = NotFound desc = could not find container \"202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce\": container with ID starting with 202b50226ec51954389e4063707f5161d10111435a45ac2d31c9d1749ae4c8ce not found: ID does not exist" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.522611 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/562073c1-2475-4e08-a121-b69cf751fa63-config-data\") pod \"562073c1-2475-4e08-a121-b69cf751fa63\" (UID: \"562073c1-2475-4e08-a121-b69cf751fa63\") " Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.522811 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz9z2\" (UniqueName: \"kubernetes.io/projected/562073c1-2475-4e08-a121-b69cf751fa63-kube-api-access-vz9z2\") pod \"562073c1-2475-4e08-a121-b69cf751fa63\" (UID: \"562073c1-2475-4e08-a121-b69cf751fa63\") " Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.522965 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562073c1-2475-4e08-a121-b69cf751fa63-combined-ca-bundle\") pod \"562073c1-2475-4e08-a121-b69cf751fa63\" (UID: \"562073c1-2475-4e08-a121-b69cf751fa63\") " Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.561364 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/562073c1-2475-4e08-a121-b69cf751fa63-kube-api-access-vz9z2" (OuterVolumeSpecName: "kube-api-access-vz9z2") pod "562073c1-2475-4e08-a121-b69cf751fa63" (UID: "562073c1-2475-4e08-a121-b69cf751fa63"). InnerVolumeSpecName "kube-api-access-vz9z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.654650 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/562073c1-2475-4e08-a121-b69cf751fa63-config-data" (OuterVolumeSpecName: "config-data") pod "562073c1-2475-4e08-a121-b69cf751fa63" (UID: "562073c1-2475-4e08-a121-b69cf751fa63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.656158 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/562073c1-2475-4e08-a121-b69cf751fa63-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.656197 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz9z2\" (UniqueName: \"kubernetes.io/projected/562073c1-2475-4e08-a121-b69cf751fa63-kube-api-access-vz9z2\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.718195 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/562073c1-2475-4e08-a121-b69cf751fa63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "562073c1-2475-4e08-a121-b69cf751fa63" (UID: "562073c1-2475-4e08-a121-b69cf751fa63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.736170 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx"] Dec 01 21:57:52 crc kubenswrapper[4857]: E1201 21:57:52.737166 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="562073c1-2475-4e08-a121-b69cf751fa63" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.737199 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="562073c1-2475-4e08-a121-b69cf751fa63" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.737478 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="562073c1-2475-4e08-a121-b69cf751fa63" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.755382 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.759069 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx"] Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.767701 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562073c1-2475-4e08-a121-b69cf751fa63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.870262 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6brv5\" (UniqueName: \"kubernetes.io/projected/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-kube-api-access-6brv5\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.870389 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.870410 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.870452 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.870525 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.870627 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-config\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.972922 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-config\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.973388 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6brv5\" (UniqueName: \"kubernetes.io/projected/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-kube-api-access-6brv5\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.973438 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.973458 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.973483 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.973519 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.973955 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-config\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.974333 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.974496 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.974629 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:52 crc kubenswrapper[4857]: I1201 21:57:52.974875 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.000673 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6brv5\" (UniqueName: \"kubernetes.io/projected/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-kube-api-access-6brv5\") pod \"dnsmasq-dns-6b7bbf7cf9-cc9tx\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.085014 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.097449 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.101736 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.108793 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.110496 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.113266 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.113358 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.113578 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.130836 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.175985 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.176198 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.176243 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.176263 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.176297 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grnml\" (UniqueName: \"kubernetes.io/projected/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-kube-api-access-grnml\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.277800 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.278243 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.278271 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.278291 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.278327 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grnml\" (UniqueName: \"kubernetes.io/projected/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-kube-api-access-grnml\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.287217 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.287291 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.291875 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.296751 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.303273 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grnml\" (UniqueName: \"kubernetes.io/projected/e28cb1b9-5dff-4741-ac99-f72b10c0c09d-kube-api-access-grnml\") pod \"nova-cell1-novncproxy-0\" (UID: \"e28cb1b9-5dff-4741-ac99-f72b10c0c09d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.512194 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.658408 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx"] Dec 01 21:57:53 crc kubenswrapper[4857]: W1201 21:57:53.670261 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3325d8c6_37b4_40e4_819a_8dff28a9c6c9.slice/crio-bf39ecdfcb76014e08a7fa23067c20b156106b06d832f5e0b1b6b59ea4b23395 WatchSource:0}: Error finding container bf39ecdfcb76014e08a7fa23067c20b156106b06d832f5e0b1b6b59ea4b23395: Status 404 returned error can't find the container with id bf39ecdfcb76014e08a7fa23067c20b156106b06d832f5e0b1b6b59ea4b23395 Dec 01 21:57:53 crc kubenswrapper[4857]: I1201 21:57:53.861639 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="562073c1-2475-4e08-a121-b69cf751fa63" path="/var/lib/kubelet/pods/562073c1-2475-4e08-a121-b69cf751fa63/volumes" Dec 01 21:57:54 crc kubenswrapper[4857]: I1201 21:57:54.001933 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 21:57:54 crc kubenswrapper[4857]: W1201 21:57:54.007577 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode28cb1b9_5dff_4741_ac99_f72b10c0c09d.slice/crio-b8e38c8f2c3de999c2d1ac2a30fe53925ca5f972fc376d289c850e451d65800a WatchSource:0}: Error finding container b8e38c8f2c3de999c2d1ac2a30fe53925ca5f972fc376d289c850e451d65800a: Status 404 returned error can't find the container with id b8e38c8f2c3de999c2d1ac2a30fe53925ca5f972fc376d289c850e451d65800a Dec 01 21:57:54 crc kubenswrapper[4857]: I1201 21:57:54.437627 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e28cb1b9-5dff-4741-ac99-f72b10c0c09d","Type":"ContainerStarted","Data":"9590faaedfcd08141c2bf5f3da3f3794338417d5538ab9ffab1aa3e7c5545bb6"} Dec 01 21:57:54 crc kubenswrapper[4857]: I1201 21:57:54.438083 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e28cb1b9-5dff-4741-ac99-f72b10c0c09d","Type":"ContainerStarted","Data":"b8e38c8f2c3de999c2d1ac2a30fe53925ca5f972fc376d289c850e451d65800a"} Dec 01 21:57:54 crc kubenswrapper[4857]: I1201 21:57:54.445444 4857 generic.go:334] "Generic (PLEG): container finished" podID="3325d8c6-37b4-40e4-819a-8dff28a9c6c9" containerID="6555feb690a2957e02e1bec2d9a2c09ba4d6fea69f4339c749c18aaab07d7de5" exitCode=0 Dec 01 21:57:54 crc kubenswrapper[4857]: I1201 21:57:54.447610 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" event={"ID":"3325d8c6-37b4-40e4-819a-8dff28a9c6c9","Type":"ContainerDied","Data":"6555feb690a2957e02e1bec2d9a2c09ba4d6fea69f4339c749c18aaab07d7de5"} Dec 01 21:57:54 crc kubenswrapper[4857]: I1201 21:57:54.447720 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" event={"ID":"3325d8c6-37b4-40e4-819a-8dff28a9c6c9","Type":"ContainerStarted","Data":"bf39ecdfcb76014e08a7fa23067c20b156106b06d832f5e0b1b6b59ea4b23395"} Dec 01 21:57:54 crc kubenswrapper[4857]: I1201 21:57:54.501897 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.5018541349999999 podStartE2EDuration="1.501854135s" podCreationTimestamp="2025-12-01 21:57:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:57:54.467798591 +0000 UTC m=+1392.957860908" watchObservedRunningTime="2025-12-01 21:57:54.501854135 +0000 UTC m=+1392.991916452" Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.100770 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.101498 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="ceilometer-central-agent" containerID="cri-o://141bc249b23cc089bad1b37bf8cfb2dbe70492872417fc7f44607080c7a6326c" gracePeriod=30 Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.102129 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="proxy-httpd" containerID="cri-o://888eed7f0dfc676d49e1922a13d2154d16c323976192546fb6ec67eb083a9dc7" gracePeriod=30 Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.102190 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="sg-core" containerID="cri-o://88e00e4e163d08859215bbed9e0f01ff178768407accda52fd4e176b8362fe31" gracePeriod=30 Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.102218 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="ceilometer-notification-agent" containerID="cri-o://cb1d06839a70f4d244af67fafd625326bc4e53c6cc8e693fad096bb8ae347a32" gracePeriod=30 Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.286846 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.475308 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" event={"ID":"3325d8c6-37b4-40e4-819a-8dff28a9c6c9","Type":"ContainerStarted","Data":"c5201045f96c5f52aacbe88de9e20041a3180be209abf5cf0d9f83ee7d48d100"} Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.475470 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.480336 4857 generic.go:334] "Generic (PLEG): container finished" podID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerID="888eed7f0dfc676d49e1922a13d2154d16c323976192546fb6ec67eb083a9dc7" exitCode=0 Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.480471 4857 generic.go:334] "Generic (PLEG): container finished" podID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerID="88e00e4e163d08859215bbed9e0f01ff178768407accda52fd4e176b8362fe31" exitCode=2 Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.480414 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bad773e6-67bc-4260-bb1a-f56ab753c7fc","Type":"ContainerDied","Data":"888eed7f0dfc676d49e1922a13d2154d16c323976192546fb6ec67eb083a9dc7"} Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.481525 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bad773e6-67bc-4260-bb1a-f56ab753c7fc","Type":"ContainerDied","Data":"88e00e4e163d08859215bbed9e0f01ff178768407accda52fd4e176b8362fe31"} Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.481706 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" containerName="nova-api-log" containerID="cri-o://21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be" gracePeriod=30 Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.481871 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" containerName="nova-api-api" containerID="cri-o://5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c" gracePeriod=30 Dec 01 21:57:55 crc kubenswrapper[4857]: I1201 21:57:55.503753 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" podStartSLOduration=3.503723565 podStartE2EDuration="3.503723565s" podCreationTimestamp="2025-12-01 21:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:57:55.494153748 +0000 UTC m=+1393.984216065" watchObservedRunningTime="2025-12-01 21:57:55.503723565 +0000 UTC m=+1393.993785872" Dec 01 21:57:56 crc kubenswrapper[4857]: I1201 21:57:56.490667 4857 generic.go:334] "Generic (PLEG): container finished" podID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" containerID="21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be" exitCode=143 Dec 01 21:57:56 crc kubenswrapper[4857]: I1201 21:57:56.491116 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eb9754b2-56a5-4a72-a3fa-3d6c280027f4","Type":"ContainerDied","Data":"21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be"} Dec 01 21:57:56 crc kubenswrapper[4857]: I1201 21:57:56.495276 4857 generic.go:334] "Generic (PLEG): container finished" podID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerID="cb1d06839a70f4d244af67fafd625326bc4e53c6cc8e693fad096bb8ae347a32" exitCode=0 Dec 01 21:57:56 crc kubenswrapper[4857]: I1201 21:57:56.495311 4857 generic.go:334] "Generic (PLEG): container finished" podID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerID="141bc249b23cc089bad1b37bf8cfb2dbe70492872417fc7f44607080c7a6326c" exitCode=0 Dec 01 21:57:56 crc kubenswrapper[4857]: I1201 21:57:56.495360 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bad773e6-67bc-4260-bb1a-f56ab753c7fc","Type":"ContainerDied","Data":"cb1d06839a70f4d244af67fafd625326bc4e53c6cc8e693fad096bb8ae347a32"} Dec 01 21:57:56 crc kubenswrapper[4857]: I1201 21:57:56.495401 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bad773e6-67bc-4260-bb1a-f56ab753c7fc","Type":"ContainerDied","Data":"141bc249b23cc089bad1b37bf8cfb2dbe70492872417fc7f44607080c7a6326c"} Dec 01 21:57:56 crc kubenswrapper[4857]: I1201 21:57:56.948191 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.131987 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bad773e6-67bc-4260-bb1a-f56ab753c7fc-run-httpd\") pod \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.132086 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgxxr\" (UniqueName: \"kubernetes.io/projected/bad773e6-67bc-4260-bb1a-f56ab753c7fc-kube-api-access-pgxxr\") pod \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.132124 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-config-data\") pod \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.132236 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-scripts\") pod \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.132274 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-combined-ca-bundle\") pod \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.132314 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-sg-core-conf-yaml\") pod \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.132356 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bad773e6-67bc-4260-bb1a-f56ab753c7fc-log-httpd\") pod \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\" (UID: \"bad773e6-67bc-4260-bb1a-f56ab753c7fc\") " Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.132494 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bad773e6-67bc-4260-bb1a-f56ab753c7fc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bad773e6-67bc-4260-bb1a-f56ab753c7fc" (UID: "bad773e6-67bc-4260-bb1a-f56ab753c7fc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.132802 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bad773e6-67bc-4260-bb1a-f56ab753c7fc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bad773e6-67bc-4260-bb1a-f56ab753c7fc" (UID: "bad773e6-67bc-4260-bb1a-f56ab753c7fc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.132970 4857 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bad773e6-67bc-4260-bb1a-f56ab753c7fc-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.132992 4857 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bad773e6-67bc-4260-bb1a-f56ab753c7fc-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.151066 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bad773e6-67bc-4260-bb1a-f56ab753c7fc-kube-api-access-pgxxr" (OuterVolumeSpecName: "kube-api-access-pgxxr") pod "bad773e6-67bc-4260-bb1a-f56ab753c7fc" (UID: "bad773e6-67bc-4260-bb1a-f56ab753c7fc"). InnerVolumeSpecName "kube-api-access-pgxxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.155673 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-scripts" (OuterVolumeSpecName: "scripts") pod "bad773e6-67bc-4260-bb1a-f56ab753c7fc" (UID: "bad773e6-67bc-4260-bb1a-f56ab753c7fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.165522 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bad773e6-67bc-4260-bb1a-f56ab753c7fc" (UID: "bad773e6-67bc-4260-bb1a-f56ab753c7fc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.218922 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bad773e6-67bc-4260-bb1a-f56ab753c7fc" (UID: "bad773e6-67bc-4260-bb1a-f56ab753c7fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.237667 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgxxr\" (UniqueName: \"kubernetes.io/projected/bad773e6-67bc-4260-bb1a-f56ab753c7fc-kube-api-access-pgxxr\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.237721 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.237742 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.237760 4857 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.257112 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-config-data" (OuterVolumeSpecName: "config-data") pod "bad773e6-67bc-4260-bb1a-f56ab753c7fc" (UID: "bad773e6-67bc-4260-bb1a-f56ab753c7fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.345666 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad773e6-67bc-4260-bb1a-f56ab753c7fc-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.509066 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bad773e6-67bc-4260-bb1a-f56ab753c7fc","Type":"ContainerDied","Data":"7ab4aac76c05dda4a1aa7c9371054ddedce2f75d64556d96d3f0fc2039182abf"} Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.509125 4857 scope.go:117] "RemoveContainer" containerID="888eed7f0dfc676d49e1922a13d2154d16c323976192546fb6ec67eb083a9dc7" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.509201 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.554459 4857 scope.go:117] "RemoveContainer" containerID="88e00e4e163d08859215bbed9e0f01ff178768407accda52fd4e176b8362fe31" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.557600 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.565271 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.600546 4857 scope.go:117] "RemoveContainer" containerID="cb1d06839a70f4d244af67fafd625326bc4e53c6cc8e693fad096bb8ae347a32" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.610002 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:57:57 crc kubenswrapper[4857]: E1201 21:57:57.610597 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="ceilometer-central-agent" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.610617 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="ceilometer-central-agent" Dec 01 21:57:57 crc kubenswrapper[4857]: E1201 21:57:57.610630 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="ceilometer-notification-agent" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.610637 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="ceilometer-notification-agent" Dec 01 21:57:57 crc kubenswrapper[4857]: E1201 21:57:57.610658 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="sg-core" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.610665 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="sg-core" Dec 01 21:57:57 crc kubenswrapper[4857]: E1201 21:57:57.610687 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="proxy-httpd" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.610692 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="proxy-httpd" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.610860 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="proxy-httpd" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.610880 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="sg-core" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.610892 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="ceilometer-central-agent" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.610900 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" containerName="ceilometer-notification-agent" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.614501 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.617557 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.621440 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.650218 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.652634 4857 scope.go:117] "RemoveContainer" containerID="141bc249b23cc089bad1b37bf8cfb2dbe70492872417fc7f44607080c7a6326c" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.653235 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.653344 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjz45\" (UniqueName: \"kubernetes.io/projected/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-kube-api-access-qjz45\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.653386 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-run-httpd\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.653428 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.653587 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-config-data\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.653699 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-scripts\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.653840 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-log-httpd\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.757638 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjz45\" (UniqueName: \"kubernetes.io/projected/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-kube-api-access-qjz45\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.757747 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-run-httpd\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.757943 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.757981 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-config-data\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.758010 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-scripts\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.758048 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-log-httpd\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.758111 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.759391 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-run-httpd\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.759475 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-log-httpd\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.763688 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.763779 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-scripts\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.765532 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-config-data\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.774015 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.786713 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjz45\" (UniqueName: \"kubernetes.io/projected/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-kube-api-access-qjz45\") pod \"ceilometer-0\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " pod="openstack/ceilometer-0" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.845374 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bad773e6-67bc-4260-bb1a-f56ab753c7fc" path="/var/lib/kubelet/pods/bad773e6-67bc-4260-bb1a-f56ab753c7fc/volumes" Dec 01 21:57:57 crc kubenswrapper[4857]: I1201 21:57:57.946993 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:57:58 crc kubenswrapper[4857]: W1201 21:57:58.464284 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfbabfb2_c3fd_4242_8404_e32e6c3acf2f.slice/crio-5dabceb8e86ffdb2ab5ca8157e38fd8eb424d59a7b9bcc504d1670f29fc268c4 WatchSource:0}: Error finding container 5dabceb8e86ffdb2ab5ca8157e38fd8eb424d59a7b9bcc504d1670f29fc268c4: Status 404 returned error can't find the container with id 5dabceb8e86ffdb2ab5ca8157e38fd8eb424d59a7b9bcc504d1670f29fc268c4 Dec 01 21:57:58 crc kubenswrapper[4857]: I1201 21:57:58.466096 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:57:58 crc kubenswrapper[4857]: I1201 21:57:58.512717 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:57:58 crc kubenswrapper[4857]: I1201 21:57:58.522425 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f","Type":"ContainerStarted","Data":"5dabceb8e86ffdb2ab5ca8157e38fd8eb424d59a7b9bcc504d1670f29fc268c4"} Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.110583 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.194601 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-logs\") pod \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.195244 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-logs" (OuterVolumeSpecName: "logs") pod "eb9754b2-56a5-4a72-a3fa-3d6c280027f4" (UID: "eb9754b2-56a5-4a72-a3fa-3d6c280027f4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.195371 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4wfb\" (UniqueName: \"kubernetes.io/projected/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-kube-api-access-j4wfb\") pod \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.195410 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-combined-ca-bundle\") pod \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.195673 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-config-data\") pod \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\" (UID: \"eb9754b2-56a5-4a72-a3fa-3d6c280027f4\") " Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.196374 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.209646 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-kube-api-access-j4wfb" (OuterVolumeSpecName: "kube-api-access-j4wfb") pod "eb9754b2-56a5-4a72-a3fa-3d6c280027f4" (UID: "eb9754b2-56a5-4a72-a3fa-3d6c280027f4"). InnerVolumeSpecName "kube-api-access-j4wfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.247058 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb9754b2-56a5-4a72-a3fa-3d6c280027f4" (UID: "eb9754b2-56a5-4a72-a3fa-3d6c280027f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.280372 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-config-data" (OuterVolumeSpecName: "config-data") pod "eb9754b2-56a5-4a72-a3fa-3d6c280027f4" (UID: "eb9754b2-56a5-4a72-a3fa-3d6c280027f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.298806 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4wfb\" (UniqueName: \"kubernetes.io/projected/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-kube-api-access-j4wfb\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.298855 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.298870 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb9754b2-56a5-4a72-a3fa-3d6c280027f4-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.531646 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f","Type":"ContainerStarted","Data":"650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef"} Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.533800 4857 generic.go:334] "Generic (PLEG): container finished" podID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" containerID="5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c" exitCode=0 Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.533845 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.533842 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eb9754b2-56a5-4a72-a3fa-3d6c280027f4","Type":"ContainerDied","Data":"5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c"} Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.538392 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eb9754b2-56a5-4a72-a3fa-3d6c280027f4","Type":"ContainerDied","Data":"92ba9cd6d5a36a000bef2f2922ac9c43fc3ce5609c1f179dbdef059c5f0bfd06"} Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.538418 4857 scope.go:117] "RemoveContainer" containerID="5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.570600 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.571598 4857 scope.go:117] "RemoveContainer" containerID="21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.586260 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.596649 4857 scope.go:117] "RemoveContainer" containerID="5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c" Dec 01 21:57:59 crc kubenswrapper[4857]: E1201 21:57:59.596975 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c\": container with ID starting with 5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c not found: ID does not exist" containerID="5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.597012 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c"} err="failed to get container status \"5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c\": rpc error: code = NotFound desc = could not find container \"5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c\": container with ID starting with 5dfdc8493fce5b5df8fdd428d3ce87a117ba8dbb967e068c30a020b3386adf6c not found: ID does not exist" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.597037 4857 scope.go:117] "RemoveContainer" containerID="21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be" Dec 01 21:57:59 crc kubenswrapper[4857]: E1201 21:57:59.597327 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be\": container with ID starting with 21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be not found: ID does not exist" containerID="21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.597356 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be"} err="failed to get container status \"21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be\": rpc error: code = NotFound desc = could not find container \"21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be\": container with ID starting with 21db9767677ea98b6fcefcc3c9f42e00cff6b199c11820c7ce39064eb5d568be not found: ID does not exist" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.608251 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:59 crc kubenswrapper[4857]: E1201 21:57:59.608827 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" containerName="nova-api-api" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.608851 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" containerName="nova-api-api" Dec 01 21:57:59 crc kubenswrapper[4857]: E1201 21:57:59.608876 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" containerName="nova-api-log" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.608887 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" containerName="nova-api-log" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.609386 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" containerName="nova-api-api" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.609449 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" containerName="nova-api-log" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.612006 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.619535 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.622254 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.622384 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.623195 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.706325 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cngd5\" (UniqueName: \"kubernetes.io/projected/9a7f9045-741f-4648-a9d5-8797c8415f7e-kube-api-access-cngd5\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.706399 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-public-tls-certs\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.706576 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.706678 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-config-data\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.706763 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a7f9045-741f-4648-a9d5-8797c8415f7e-logs\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.706881 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.809169 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.809263 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cngd5\" (UniqueName: \"kubernetes.io/projected/9a7f9045-741f-4648-a9d5-8797c8415f7e-kube-api-access-cngd5\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.809342 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-public-tls-certs\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.809398 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.809438 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-config-data\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.809474 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a7f9045-741f-4648-a9d5-8797c8415f7e-logs\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.810024 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a7f9045-741f-4648-a9d5-8797c8415f7e-logs\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.815572 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-public-tls-certs\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.816263 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.816742 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.820829 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-config-data\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.842117 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cngd5\" (UniqueName: \"kubernetes.io/projected/9a7f9045-741f-4648-a9d5-8797c8415f7e-kube-api-access-cngd5\") pod \"nova-api-0\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " pod="openstack/nova-api-0" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.845761 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb9754b2-56a5-4a72-a3fa-3d6c280027f4" path="/var/lib/kubelet/pods/eb9754b2-56a5-4a72-a3fa-3d6c280027f4/volumes" Dec 01 21:57:59 crc kubenswrapper[4857]: I1201 21:57:59.943071 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:58:00 crc kubenswrapper[4857]: I1201 21:58:00.500770 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:58:00 crc kubenswrapper[4857]: I1201 21:58:00.559608 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f","Type":"ContainerStarted","Data":"d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c"} Dec 01 21:58:00 crc kubenswrapper[4857]: I1201 21:58:00.561439 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9a7f9045-741f-4648-a9d5-8797c8415f7e","Type":"ContainerStarted","Data":"c3dee29678a50a18d686d1dc90c43768245aa8e94411c8ac18777ca0d14dc457"} Dec 01 21:58:01 crc kubenswrapper[4857]: I1201 21:58:01.588372 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f","Type":"ContainerStarted","Data":"3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba"} Dec 01 21:58:01 crc kubenswrapper[4857]: I1201 21:58:01.591668 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9a7f9045-741f-4648-a9d5-8797c8415f7e","Type":"ContainerStarted","Data":"ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa"} Dec 01 21:58:01 crc kubenswrapper[4857]: I1201 21:58:01.591730 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9a7f9045-741f-4648-a9d5-8797c8415f7e","Type":"ContainerStarted","Data":"ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76"} Dec 01 21:58:01 crc kubenswrapper[4857]: I1201 21:58:01.632663 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.632614725 podStartE2EDuration="2.632614725s" podCreationTimestamp="2025-12-01 21:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:58:01.613969353 +0000 UTC m=+1400.104031710" watchObservedRunningTime="2025-12-01 21:58:01.632614725 +0000 UTC m=+1400.122677052" Dec 01 21:58:02 crc kubenswrapper[4857]: I1201 21:58:02.605462 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f","Type":"ContainerStarted","Data":"304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31"} Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.103403 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.347198 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.650927399 podStartE2EDuration="6.347168501s" podCreationTimestamp="2025-12-01 21:57:57 +0000 UTC" firstStartedPulling="2025-12-01 21:57:58.468874445 +0000 UTC m=+1396.958936802" lastFinishedPulling="2025-12-01 21:58:02.165115587 +0000 UTC m=+1400.655177904" observedRunningTime="2025-12-01 21:58:02.629202175 +0000 UTC m=+1401.119264492" watchObservedRunningTime="2025-12-01 21:58:03.347168501 +0000 UTC m=+1401.837230818" Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.423124 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-fl6ws"] Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.423435 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" podUID="fc6ab107-b310-4f97-8bef-695b1a75f728" containerName="dnsmasq-dns" containerID="cri-o://c5fe7d8f3d0454204f851dc543dcccd87f04767356e8108a8c524c3a66dec83b" gracePeriod=10 Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.512489 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.547651 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.618868 4857 generic.go:334] "Generic (PLEG): container finished" podID="fc6ab107-b310-4f97-8bef-695b1a75f728" containerID="c5fe7d8f3d0454204f851dc543dcccd87f04767356e8108a8c524c3a66dec83b" exitCode=0 Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.620622 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" event={"ID":"fc6ab107-b310-4f97-8bef-695b1a75f728","Type":"ContainerDied","Data":"c5fe7d8f3d0454204f851dc543dcccd87f04767356e8108a8c524c3a66dec83b"} Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.621210 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.647936 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.990102 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-lwhqc"] Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.991713 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.998416 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 01 21:58:03 crc kubenswrapper[4857]: I1201 21:58:03.998965 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.025852 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lwhqc"] Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.179615 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-config-data\") pod \"nova-cell1-cell-mapping-lwhqc\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.179716 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mc8m\" (UniqueName: \"kubernetes.io/projected/345660e3-c8c5-45ef-a5e5-315d7254a41c-kube-api-access-7mc8m\") pod \"nova-cell1-cell-mapping-lwhqc\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.179769 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lwhqc\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.179838 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-scripts\") pod \"nova-cell1-cell-mapping-lwhqc\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.200192 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.281609 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-config-data\") pod \"nova-cell1-cell-mapping-lwhqc\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.281668 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mc8m\" (UniqueName: \"kubernetes.io/projected/345660e3-c8c5-45ef-a5e5-315d7254a41c-kube-api-access-7mc8m\") pod \"nova-cell1-cell-mapping-lwhqc\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.281706 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lwhqc\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.281740 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-scripts\") pod \"nova-cell1-cell-mapping-lwhqc\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.289937 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lwhqc\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.298753 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-config-data\") pod \"nova-cell1-cell-mapping-lwhqc\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.308098 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-scripts\") pod \"nova-cell1-cell-mapping-lwhqc\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.324909 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mc8m\" (UniqueName: \"kubernetes.io/projected/345660e3-c8c5-45ef-a5e5-315d7254a41c-kube-api-access-7mc8m\") pod \"nova-cell1-cell-mapping-lwhqc\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.357443 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.383884 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-dns-svc\") pod \"fc6ab107-b310-4f97-8bef-695b1a75f728\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.383987 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwsr2\" (UniqueName: \"kubernetes.io/projected/fc6ab107-b310-4f97-8bef-695b1a75f728-kube-api-access-mwsr2\") pod \"fc6ab107-b310-4f97-8bef-695b1a75f728\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.384027 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-dns-swift-storage-0\") pod \"fc6ab107-b310-4f97-8bef-695b1a75f728\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.384087 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-config\") pod \"fc6ab107-b310-4f97-8bef-695b1a75f728\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.384146 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-ovsdbserver-sb\") pod \"fc6ab107-b310-4f97-8bef-695b1a75f728\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.384197 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-ovsdbserver-nb\") pod \"fc6ab107-b310-4f97-8bef-695b1a75f728\" (UID: \"fc6ab107-b310-4f97-8bef-695b1a75f728\") " Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.399197 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc6ab107-b310-4f97-8bef-695b1a75f728-kube-api-access-mwsr2" (OuterVolumeSpecName: "kube-api-access-mwsr2") pod "fc6ab107-b310-4f97-8bef-695b1a75f728" (UID: "fc6ab107-b310-4f97-8bef-695b1a75f728"). InnerVolumeSpecName "kube-api-access-mwsr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.451699 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fc6ab107-b310-4f97-8bef-695b1a75f728" (UID: "fc6ab107-b310-4f97-8bef-695b1a75f728"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.452621 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fc6ab107-b310-4f97-8bef-695b1a75f728" (UID: "fc6ab107-b310-4f97-8bef-695b1a75f728"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.467087 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fc6ab107-b310-4f97-8bef-695b1a75f728" (UID: "fc6ab107-b310-4f97-8bef-695b1a75f728"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.474196 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fc6ab107-b310-4f97-8bef-695b1a75f728" (UID: "fc6ab107-b310-4f97-8bef-695b1a75f728"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.481522 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-config" (OuterVolumeSpecName: "config") pod "fc6ab107-b310-4f97-8bef-695b1a75f728" (UID: "fc6ab107-b310-4f97-8bef-695b1a75f728"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.487211 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.487259 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwsr2\" (UniqueName: \"kubernetes.io/projected/fc6ab107-b310-4f97-8bef-695b1a75f728-kube-api-access-mwsr2\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.487279 4857 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.487291 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.487303 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.487314 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc6ab107-b310-4f97-8bef-695b1a75f728-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.637315 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" event={"ID":"fc6ab107-b310-4f97-8bef-695b1a75f728","Type":"ContainerDied","Data":"27e5949444e915d4996447210ab491a80f406be7138117caf1f2fd73c52fc255"} Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.637424 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-fl6ws" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.637434 4857 scope.go:117] "RemoveContainer" containerID="c5fe7d8f3d0454204f851dc543dcccd87f04767356e8108a8c524c3a66dec83b" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.683108 4857 scope.go:117] "RemoveContainer" containerID="e4fce0baf3b19bf0c6fd344c73551228cb88e80e1298c2b019f82171e1b3d532" Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.692932 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-fl6ws"] Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.704534 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-fl6ws"] Dec 01 21:58:04 crc kubenswrapper[4857]: I1201 21:58:04.886104 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lwhqc"] Dec 01 21:58:05 crc kubenswrapper[4857]: I1201 21:58:05.650315 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lwhqc" event={"ID":"345660e3-c8c5-45ef-a5e5-315d7254a41c","Type":"ContainerStarted","Data":"180c474545583e5dd7bd2a9be4851eaa2bccaf82f0383d261a55dedee7589fad"} Dec 01 21:58:05 crc kubenswrapper[4857]: I1201 21:58:05.650700 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lwhqc" event={"ID":"345660e3-c8c5-45ef-a5e5-315d7254a41c","Type":"ContainerStarted","Data":"fa87ef5ea4dfd4391fbd2b377a4e3a4033b04ee9a1edeebbe3eba9183d450088"} Dec 01 21:58:05 crc kubenswrapper[4857]: I1201 21:58:05.668110 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-lwhqc" podStartSLOduration=2.668091601 podStartE2EDuration="2.668091601s" podCreationTimestamp="2025-12-01 21:58:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:58:05.664709787 +0000 UTC m=+1404.154772094" watchObservedRunningTime="2025-12-01 21:58:05.668091601 +0000 UTC m=+1404.158153918" Dec 01 21:58:05 crc kubenswrapper[4857]: I1201 21:58:05.857458 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc6ab107-b310-4f97-8bef-695b1a75f728" path="/var/lib/kubelet/pods/fc6ab107-b310-4f97-8bef-695b1a75f728/volumes" Dec 01 21:58:09 crc kubenswrapper[4857]: I1201 21:58:09.943208 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 21:58:09 crc kubenswrapper[4857]: I1201 21:58:09.944283 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 21:58:10 crc kubenswrapper[4857]: I1201 21:58:10.713424 4857 generic.go:334] "Generic (PLEG): container finished" podID="345660e3-c8c5-45ef-a5e5-315d7254a41c" containerID="180c474545583e5dd7bd2a9be4851eaa2bccaf82f0383d261a55dedee7589fad" exitCode=0 Dec 01 21:58:10 crc kubenswrapper[4857]: I1201 21:58:10.713482 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lwhqc" event={"ID":"345660e3-c8c5-45ef-a5e5-315d7254a41c","Type":"ContainerDied","Data":"180c474545583e5dd7bd2a9be4851eaa2bccaf82f0383d261a55dedee7589fad"} Dec 01 21:58:10 crc kubenswrapper[4857]: I1201 21:58:10.957233 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9a7f9045-741f-4648-a9d5-8797c8415f7e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 21:58:10 crc kubenswrapper[4857]: I1201 21:58:10.957490 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9a7f9045-741f-4648-a9d5-8797c8415f7e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.227870 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.372532 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-combined-ca-bundle\") pod \"345660e3-c8c5-45ef-a5e5-315d7254a41c\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.372752 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mc8m\" (UniqueName: \"kubernetes.io/projected/345660e3-c8c5-45ef-a5e5-315d7254a41c-kube-api-access-7mc8m\") pod \"345660e3-c8c5-45ef-a5e5-315d7254a41c\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.372789 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-scripts\") pod \"345660e3-c8c5-45ef-a5e5-315d7254a41c\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.372915 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-config-data\") pod \"345660e3-c8c5-45ef-a5e5-315d7254a41c\" (UID: \"345660e3-c8c5-45ef-a5e5-315d7254a41c\") " Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.380256 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-scripts" (OuterVolumeSpecName: "scripts") pod "345660e3-c8c5-45ef-a5e5-315d7254a41c" (UID: "345660e3-c8c5-45ef-a5e5-315d7254a41c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.380465 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/345660e3-c8c5-45ef-a5e5-315d7254a41c-kube-api-access-7mc8m" (OuterVolumeSpecName: "kube-api-access-7mc8m") pod "345660e3-c8c5-45ef-a5e5-315d7254a41c" (UID: "345660e3-c8c5-45ef-a5e5-315d7254a41c"). InnerVolumeSpecName "kube-api-access-7mc8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.409509 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-config-data" (OuterVolumeSpecName: "config-data") pod "345660e3-c8c5-45ef-a5e5-315d7254a41c" (UID: "345660e3-c8c5-45ef-a5e5-315d7254a41c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.411985 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "345660e3-c8c5-45ef-a5e5-315d7254a41c" (UID: "345660e3-c8c5-45ef-a5e5-315d7254a41c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.475748 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.475796 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.475811 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mc8m\" (UniqueName: \"kubernetes.io/projected/345660e3-c8c5-45ef-a5e5-315d7254a41c-kube-api-access-7mc8m\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.475821 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345660e3-c8c5-45ef-a5e5-315d7254a41c-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.739897 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lwhqc" event={"ID":"345660e3-c8c5-45ef-a5e5-315d7254a41c","Type":"ContainerDied","Data":"fa87ef5ea4dfd4391fbd2b377a4e3a4033b04ee9a1edeebbe3eba9183d450088"} Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.740147 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa87ef5ea4dfd4391fbd2b377a4e3a4033b04ee9a1edeebbe3eba9183d450088" Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.740363 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lwhqc" Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.840671 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.840907 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9a7f9045-741f-4648-a9d5-8797c8415f7e" containerName="nova-api-log" containerID="cri-o://ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76" gracePeriod=30 Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.841063 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9a7f9045-741f-4648-a9d5-8797c8415f7e" containerName="nova-api-api" containerID="cri-o://ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa" gracePeriod=30 Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.890090 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.892612 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c402ef80-d243-4a46-bd8f-9fa0036d771e" containerName="nova-scheduler-scheduler" containerID="cri-o://3931aa3a7a90531868722bb8b84722451b0a11cc16dd4fdc24674887903a3cd3" gracePeriod=30 Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.916868 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.917483 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerName="nova-metadata-log" containerID="cri-o://8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664" gracePeriod=30 Dec 01 21:58:12 crc kubenswrapper[4857]: I1201 21:58:12.917550 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerName="nova-metadata-metadata" containerID="cri-o://911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f" gracePeriod=30 Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.423547 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zxcr6"] Dec 01 21:58:13 crc kubenswrapper[4857]: E1201 21:58:13.423990 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345660e3-c8c5-45ef-a5e5-315d7254a41c" containerName="nova-manage" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.424008 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="345660e3-c8c5-45ef-a5e5-315d7254a41c" containerName="nova-manage" Dec 01 21:58:13 crc kubenswrapper[4857]: E1201 21:58:13.424034 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc6ab107-b310-4f97-8bef-695b1a75f728" containerName="init" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.424056 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc6ab107-b310-4f97-8bef-695b1a75f728" containerName="init" Dec 01 21:58:13 crc kubenswrapper[4857]: E1201 21:58:13.424064 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc6ab107-b310-4f97-8bef-695b1a75f728" containerName="dnsmasq-dns" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.424071 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc6ab107-b310-4f97-8bef-695b1a75f728" containerName="dnsmasq-dns" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.424292 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="345660e3-c8c5-45ef-a5e5-315d7254a41c" containerName="nova-manage" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.424306 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc6ab107-b310-4f97-8bef-695b1a75f728" containerName="dnsmasq-dns" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.426032 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.433323 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zxcr6"] Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.597386 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ncpj\" (UniqueName: \"kubernetes.io/projected/4e928e86-c08b-4b61-acbb-4a86fa4464a6-kube-api-access-6ncpj\") pod \"redhat-operators-zxcr6\" (UID: \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\") " pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.597537 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e928e86-c08b-4b61-acbb-4a86fa4464a6-utilities\") pod \"redhat-operators-zxcr6\" (UID: \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\") " pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.597681 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e928e86-c08b-4b61-acbb-4a86fa4464a6-catalog-content\") pod \"redhat-operators-zxcr6\" (UID: \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\") " pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.699117 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e928e86-c08b-4b61-acbb-4a86fa4464a6-catalog-content\") pod \"redhat-operators-zxcr6\" (UID: \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\") " pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.699245 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ncpj\" (UniqueName: \"kubernetes.io/projected/4e928e86-c08b-4b61-acbb-4a86fa4464a6-kube-api-access-6ncpj\") pod \"redhat-operators-zxcr6\" (UID: \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\") " pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.699333 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e928e86-c08b-4b61-acbb-4a86fa4464a6-utilities\") pod \"redhat-operators-zxcr6\" (UID: \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\") " pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.699696 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e928e86-c08b-4b61-acbb-4a86fa4464a6-catalog-content\") pod \"redhat-operators-zxcr6\" (UID: \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\") " pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.699707 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e928e86-c08b-4b61-acbb-4a86fa4464a6-utilities\") pod \"redhat-operators-zxcr6\" (UID: \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\") " pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.775734 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ncpj\" (UniqueName: \"kubernetes.io/projected/4e928e86-c08b-4b61-acbb-4a86fa4464a6-kube-api-access-6ncpj\") pod \"redhat-operators-zxcr6\" (UID: \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\") " pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.780288 4857 generic.go:334] "Generic (PLEG): container finished" podID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerID="8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664" exitCode=143 Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.780350 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"14d98749-4c8f-4ad4-abc4-3229198de8f7","Type":"ContainerDied","Data":"8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664"} Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.785790 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a7f9045-741f-4648-a9d5-8797c8415f7e" containerID="ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76" exitCode=143 Dec 01 21:58:13 crc kubenswrapper[4857]: I1201 21:58:13.785839 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9a7f9045-741f-4648-a9d5-8797c8415f7e","Type":"ContainerDied","Data":"ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76"} Dec 01 21:58:14 crc kubenswrapper[4857]: I1201 21:58:14.059293 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:14 crc kubenswrapper[4857]: I1201 21:58:14.563321 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zxcr6"] Dec 01 21:58:14 crc kubenswrapper[4857]: W1201 21:58:14.570481 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e928e86_c08b_4b61_acbb_4a86fa4464a6.slice/crio-b8bf608264eca8094ed6d17c726019a788dec8e7f085b8318fe335aa2ff74760 WatchSource:0}: Error finding container b8bf608264eca8094ed6d17c726019a788dec8e7f085b8318fe335aa2ff74760: Status 404 returned error can't find the container with id b8bf608264eca8094ed6d17c726019a788dec8e7f085b8318fe335aa2ff74760 Dec 01 21:58:14 crc kubenswrapper[4857]: I1201 21:58:14.797616 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxcr6" event={"ID":"4e928e86-c08b-4b61-acbb-4a86fa4464a6","Type":"ContainerStarted","Data":"ca8babef73bb4515085c02382a7fa70fb49c68f7d9dacaac791f657746813158"} Dec 01 21:58:14 crc kubenswrapper[4857]: I1201 21:58:14.797907 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxcr6" event={"ID":"4e928e86-c08b-4b61-acbb-4a86fa4464a6","Type":"ContainerStarted","Data":"b8bf608264eca8094ed6d17c726019a788dec8e7f085b8318fe335aa2ff74760"} Dec 01 21:58:14 crc kubenswrapper[4857]: I1201 21:58:14.800566 4857 generic.go:334] "Generic (PLEG): container finished" podID="c402ef80-d243-4a46-bd8f-9fa0036d771e" containerID="3931aa3a7a90531868722bb8b84722451b0a11cc16dd4fdc24674887903a3cd3" exitCode=0 Dec 01 21:58:14 crc kubenswrapper[4857]: I1201 21:58:14.800656 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c402ef80-d243-4a46-bd8f-9fa0036d771e","Type":"ContainerDied","Data":"3931aa3a7a90531868722bb8b84722451b0a11cc16dd4fdc24674887903a3cd3"} Dec 01 21:58:14 crc kubenswrapper[4857]: I1201 21:58:14.800788 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c402ef80-d243-4a46-bd8f-9fa0036d771e","Type":"ContainerDied","Data":"3d1667229fe8bee04af3b8ac172f200d61a20f5d788e37a02e49f6be922d3a05"} Dec 01 21:58:14 crc kubenswrapper[4857]: I1201 21:58:14.800885 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d1667229fe8bee04af3b8ac172f200d61a20f5d788e37a02e49f6be922d3a05" Dec 01 21:58:14 crc kubenswrapper[4857]: I1201 21:58:14.891842 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.026001 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c402ef80-d243-4a46-bd8f-9fa0036d771e-config-data\") pod \"c402ef80-d243-4a46-bd8f-9fa0036d771e\" (UID: \"c402ef80-d243-4a46-bd8f-9fa0036d771e\") " Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.026410 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwpx4\" (UniqueName: \"kubernetes.io/projected/c402ef80-d243-4a46-bd8f-9fa0036d771e-kube-api-access-kwpx4\") pod \"c402ef80-d243-4a46-bd8f-9fa0036d771e\" (UID: \"c402ef80-d243-4a46-bd8f-9fa0036d771e\") " Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.026457 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c402ef80-d243-4a46-bd8f-9fa0036d771e-combined-ca-bundle\") pod \"c402ef80-d243-4a46-bd8f-9fa0036d771e\" (UID: \"c402ef80-d243-4a46-bd8f-9fa0036d771e\") " Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.033810 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c402ef80-d243-4a46-bd8f-9fa0036d771e-kube-api-access-kwpx4" (OuterVolumeSpecName: "kube-api-access-kwpx4") pod "c402ef80-d243-4a46-bd8f-9fa0036d771e" (UID: "c402ef80-d243-4a46-bd8f-9fa0036d771e"). InnerVolumeSpecName "kube-api-access-kwpx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.054363 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c402ef80-d243-4a46-bd8f-9fa0036d771e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c402ef80-d243-4a46-bd8f-9fa0036d771e" (UID: "c402ef80-d243-4a46-bd8f-9fa0036d771e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.062343 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c402ef80-d243-4a46-bd8f-9fa0036d771e-config-data" (OuterVolumeSpecName: "config-data") pod "c402ef80-d243-4a46-bd8f-9fa0036d771e" (UID: "c402ef80-d243-4a46-bd8f-9fa0036d771e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.129475 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwpx4\" (UniqueName: \"kubernetes.io/projected/c402ef80-d243-4a46-bd8f-9fa0036d771e-kube-api-access-kwpx4\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.129524 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c402ef80-d243-4a46-bd8f-9fa0036d771e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.129534 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c402ef80-d243-4a46-bd8f-9fa0036d771e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.810030 4857 generic.go:334] "Generic (PLEG): container finished" podID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" containerID="ca8babef73bb4515085c02382a7fa70fb49c68f7d9dacaac791f657746813158" exitCode=0 Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.810115 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.810136 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxcr6" event={"ID":"4e928e86-c08b-4b61-acbb-4a86fa4464a6","Type":"ContainerDied","Data":"ca8babef73bb4515085c02382a7fa70fb49c68f7d9dacaac791f657746813158"} Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.873815 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.888908 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.906454 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:58:15 crc kubenswrapper[4857]: E1201 21:58:15.906910 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c402ef80-d243-4a46-bd8f-9fa0036d771e" containerName="nova-scheduler-scheduler" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.906927 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c402ef80-d243-4a46-bd8f-9fa0036d771e" containerName="nova-scheduler-scheduler" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.907153 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c402ef80-d243-4a46-bd8f-9fa0036d771e" containerName="nova-scheduler-scheduler" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.907807 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.910429 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 21:58:15 crc kubenswrapper[4857]: I1201 21:58:15.920236 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.047773 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2859495a-e026-400f-a304-05e25e961f2e-config-data\") pod \"nova-scheduler-0\" (UID: \"2859495a-e026-400f-a304-05e25e961f2e\") " pod="openstack/nova-scheduler-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.047979 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2859495a-e026-400f-a304-05e25e961f2e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2859495a-e026-400f-a304-05e25e961f2e\") " pod="openstack/nova-scheduler-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.048018 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktdcz\" (UniqueName: \"kubernetes.io/projected/2859495a-e026-400f-a304-05e25e961f2e-kube-api-access-ktdcz\") pod \"nova-scheduler-0\" (UID: \"2859495a-e026-400f-a304-05e25e961f2e\") " pod="openstack/nova-scheduler-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.053019 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": read tcp 10.217.0.2:48868->10.217.0.199:8775: read: connection reset by peer" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.053444 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": read tcp 10.217.0.2:48882->10.217.0.199:8775: read: connection reset by peer" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.149835 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2859495a-e026-400f-a304-05e25e961f2e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2859495a-e026-400f-a304-05e25e961f2e\") " pod="openstack/nova-scheduler-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.149891 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktdcz\" (UniqueName: \"kubernetes.io/projected/2859495a-e026-400f-a304-05e25e961f2e-kube-api-access-ktdcz\") pod \"nova-scheduler-0\" (UID: \"2859495a-e026-400f-a304-05e25e961f2e\") " pod="openstack/nova-scheduler-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.149973 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2859495a-e026-400f-a304-05e25e961f2e-config-data\") pod \"nova-scheduler-0\" (UID: \"2859495a-e026-400f-a304-05e25e961f2e\") " pod="openstack/nova-scheduler-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.154091 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2859495a-e026-400f-a304-05e25e961f2e-config-data\") pod \"nova-scheduler-0\" (UID: \"2859495a-e026-400f-a304-05e25e961f2e\") " pod="openstack/nova-scheduler-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.154803 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2859495a-e026-400f-a304-05e25e961f2e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2859495a-e026-400f-a304-05e25e961f2e\") " pod="openstack/nova-scheduler-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.173633 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktdcz\" (UniqueName: \"kubernetes.io/projected/2859495a-e026-400f-a304-05e25e961f2e-kube-api-access-ktdcz\") pod \"nova-scheduler-0\" (UID: \"2859495a-e026-400f-a304-05e25e961f2e\") " pod="openstack/nova-scheduler-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.237180 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.562264 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.610458 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.663420 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cngd5\" (UniqueName: \"kubernetes.io/projected/9a7f9045-741f-4648-a9d5-8797c8415f7e-kube-api-access-cngd5\") pod \"9a7f9045-741f-4648-a9d5-8797c8415f7e\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.663500 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-config-data\") pod \"9a7f9045-741f-4648-a9d5-8797c8415f7e\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.664599 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-public-tls-certs\") pod \"9a7f9045-741f-4648-a9d5-8797c8415f7e\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.664831 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a7f9045-741f-4648-a9d5-8797c8415f7e-logs\") pod \"9a7f9045-741f-4648-a9d5-8797c8415f7e\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.665558 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-internal-tls-certs\") pod \"9a7f9045-741f-4648-a9d5-8797c8415f7e\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.665706 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-combined-ca-bundle\") pod \"9a7f9045-741f-4648-a9d5-8797c8415f7e\" (UID: \"9a7f9045-741f-4648-a9d5-8797c8415f7e\") " Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.666083 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a7f9045-741f-4648-a9d5-8797c8415f7e-logs" (OuterVolumeSpecName: "logs") pod "9a7f9045-741f-4648-a9d5-8797c8415f7e" (UID: "9a7f9045-741f-4648-a9d5-8797c8415f7e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.707991 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a7f9045-741f-4648-a9d5-8797c8415f7e-kube-api-access-cngd5" (OuterVolumeSpecName: "kube-api-access-cngd5") pod "9a7f9045-741f-4648-a9d5-8797c8415f7e" (UID: "9a7f9045-741f-4648-a9d5-8797c8415f7e"). InnerVolumeSpecName "kube-api-access-cngd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.709168 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cngd5\" (UniqueName: \"kubernetes.io/projected/9a7f9045-741f-4648-a9d5-8797c8415f7e-kube-api-access-cngd5\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.709230 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a7f9045-741f-4648-a9d5-8797c8415f7e-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.712995 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a7f9045-741f-4648-a9d5-8797c8415f7e" (UID: "9a7f9045-741f-4648-a9d5-8797c8415f7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.713166 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-config-data" (OuterVolumeSpecName: "config-data") pod "9a7f9045-741f-4648-a9d5-8797c8415f7e" (UID: "9a7f9045-741f-4648-a9d5-8797c8415f7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.746302 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9a7f9045-741f-4648-a9d5-8797c8415f7e" (UID: "9a7f9045-741f-4648-a9d5-8797c8415f7e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.772388 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9a7f9045-741f-4648-a9d5-8797c8415f7e" (UID: "9a7f9045-741f-4648-a9d5-8797c8415f7e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.797800 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.810369 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-nova-metadata-tls-certs\") pod \"14d98749-4c8f-4ad4-abc4-3229198de8f7\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.810533 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzqh2\" (UniqueName: \"kubernetes.io/projected/14d98749-4c8f-4ad4-abc4-3229198de8f7-kube-api-access-nzqh2\") pod \"14d98749-4c8f-4ad4-abc4-3229198de8f7\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.810706 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-config-data\") pod \"14d98749-4c8f-4ad4-abc4-3229198de8f7\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.810754 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-combined-ca-bundle\") pod \"14d98749-4c8f-4ad4-abc4-3229198de8f7\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.810774 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14d98749-4c8f-4ad4-abc4-3229198de8f7-logs\") pod \"14d98749-4c8f-4ad4-abc4-3229198de8f7\" (UID: \"14d98749-4c8f-4ad4-abc4-3229198de8f7\") " Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.811277 4857 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.811288 4857 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.811297 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.811306 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a7f9045-741f-4648-a9d5-8797c8415f7e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.812095 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14d98749-4c8f-4ad4-abc4-3229198de8f7-logs" (OuterVolumeSpecName: "logs") pod "14d98749-4c8f-4ad4-abc4-3229198de8f7" (UID: "14d98749-4c8f-4ad4-abc4-3229198de8f7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.816806 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14d98749-4c8f-4ad4-abc4-3229198de8f7-kube-api-access-nzqh2" (OuterVolumeSpecName: "kube-api-access-nzqh2") pod "14d98749-4c8f-4ad4-abc4-3229198de8f7" (UID: "14d98749-4c8f-4ad4-abc4-3229198de8f7"). InnerVolumeSpecName "kube-api-access-nzqh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.827215 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a7f9045-741f-4648-a9d5-8797c8415f7e" containerID="ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa" exitCode=0 Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.827291 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9a7f9045-741f-4648-a9d5-8797c8415f7e","Type":"ContainerDied","Data":"ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa"} Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.827322 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9a7f9045-741f-4648-a9d5-8797c8415f7e","Type":"ContainerDied","Data":"c3dee29678a50a18d686d1dc90c43768245aa8e94411c8ac18777ca0d14dc457"} Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.827338 4857 scope.go:117] "RemoveContainer" containerID="ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.827672 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.830949 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2859495a-e026-400f-a304-05e25e961f2e","Type":"ContainerStarted","Data":"a19705025aa7650eb6d85a218da198a1166567c054cb8efac640139216b22e64"} Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.832967 4857 generic.go:334] "Generic (PLEG): container finished" podID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerID="911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f" exitCode=0 Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.833005 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"14d98749-4c8f-4ad4-abc4-3229198de8f7","Type":"ContainerDied","Data":"911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f"} Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.833028 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.833031 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"14d98749-4c8f-4ad4-abc4-3229198de8f7","Type":"ContainerDied","Data":"6d9510d9fd0b63ed185c4b9eb0bcd786069d2f47b2d50353e845e8a89d36b1d3"} Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.865583 4857 scope.go:117] "RemoveContainer" containerID="ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.868384 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14d98749-4c8f-4ad4-abc4-3229198de8f7" (UID: "14d98749-4c8f-4ad4-abc4-3229198de8f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.880348 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-config-data" (OuterVolumeSpecName: "config-data") pod "14d98749-4c8f-4ad4-abc4-3229198de8f7" (UID: "14d98749-4c8f-4ad4-abc4-3229198de8f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.887235 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "14d98749-4c8f-4ad4-abc4-3229198de8f7" (UID: "14d98749-4c8f-4ad4-abc4-3229198de8f7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.920862 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.920916 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.920937 4857 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14d98749-4c8f-4ad4-abc4-3229198de8f7-logs\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.920951 4857 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/14d98749-4c8f-4ad4-abc4-3229198de8f7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.920965 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzqh2\" (UniqueName: \"kubernetes.io/projected/14d98749-4c8f-4ad4-abc4-3229198de8f7-kube-api-access-nzqh2\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:16 crc kubenswrapper[4857]: I1201 21:58:16.970684 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.004696 4857 scope.go:117] "RemoveContainer" containerID="ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.005110 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:58:17 crc kubenswrapper[4857]: E1201 21:58:17.007105 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa\": container with ID starting with ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa not found: ID does not exist" containerID="ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.007172 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa"} err="failed to get container status \"ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa\": rpc error: code = NotFound desc = could not find container \"ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa\": container with ID starting with ccdc3029c9ce2ac33a83f2d7d364a21d2855cfe0f4c702d378973d60dfdc4ffa not found: ID does not exist" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.007229 4857 scope.go:117] "RemoveContainer" containerID="ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76" Dec 01 21:58:17 crc kubenswrapper[4857]: E1201 21:58:17.007583 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76\": container with ID starting with ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76 not found: ID does not exist" containerID="ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.007625 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76"} err="failed to get container status \"ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76\": rpc error: code = NotFound desc = could not find container \"ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76\": container with ID starting with ec831e4ec8f2ce19e3cdff548f25533cfb01264166287dd21c34213a19f6de76 not found: ID does not exist" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.007646 4857 scope.go:117] "RemoveContainer" containerID="911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.021838 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 21:58:17 crc kubenswrapper[4857]: E1201 21:58:17.022457 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a7f9045-741f-4648-a9d5-8797c8415f7e" containerName="nova-api-log" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.022476 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a7f9045-741f-4648-a9d5-8797c8415f7e" containerName="nova-api-log" Dec 01 21:58:17 crc kubenswrapper[4857]: E1201 21:58:17.022490 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerName="nova-metadata-log" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.022502 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerName="nova-metadata-log" Dec 01 21:58:17 crc kubenswrapper[4857]: E1201 21:58:17.022521 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerName="nova-metadata-metadata" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.022527 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerName="nova-metadata-metadata" Dec 01 21:58:17 crc kubenswrapper[4857]: E1201 21:58:17.022535 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a7f9045-741f-4648-a9d5-8797c8415f7e" containerName="nova-api-api" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.022540 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a7f9045-741f-4648-a9d5-8797c8415f7e" containerName="nova-api-api" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.022728 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a7f9045-741f-4648-a9d5-8797c8415f7e" containerName="nova-api-log" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.022753 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a7f9045-741f-4648-a9d5-8797c8415f7e" containerName="nova-api-api" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.022769 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerName="nova-metadata-log" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.022779 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" containerName="nova-metadata-metadata" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.024885 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.027369 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.027590 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.030624 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.033806 4857 scope.go:117] "RemoveContainer" containerID="8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.037340 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.059352 4857 scope.go:117] "RemoveContainer" containerID="911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f" Dec 01 21:58:17 crc kubenswrapper[4857]: E1201 21:58:17.060952 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f\": container with ID starting with 911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f not found: ID does not exist" containerID="911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.060988 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f"} err="failed to get container status \"911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f\": rpc error: code = NotFound desc = could not find container \"911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f\": container with ID starting with 911d11f2fc2fe6234a4deea9256ef438e70a1c3cda1f692922055520f974870f not found: ID does not exist" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.061016 4857 scope.go:117] "RemoveContainer" containerID="8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664" Dec 01 21:58:17 crc kubenswrapper[4857]: E1201 21:58:17.061247 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664\": container with ID starting with 8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664 not found: ID does not exist" containerID="8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.061274 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664"} err="failed to get container status \"8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664\": rpc error: code = NotFound desc = could not find container \"8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664\": container with ID starting with 8b338a135576b5cc93b538d17aeba761d65cf84a54705bdc1050c1b30feee664 not found: ID does not exist" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.125597 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcd76a72-5869-4014-8bf7-ee57297feb85-config-data\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.125939 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dcd76a72-5869-4014-8bf7-ee57297feb85-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.126146 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dcd76a72-5869-4014-8bf7-ee57297feb85-public-tls-certs\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.126289 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcd76a72-5869-4014-8bf7-ee57297feb85-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.126445 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dcd76a72-5869-4014-8bf7-ee57297feb85-logs\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.126564 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dz58\" (UniqueName: \"kubernetes.io/projected/dcd76a72-5869-4014-8bf7-ee57297feb85-kube-api-access-2dz58\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.177020 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.187658 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.205307 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.208113 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.210491 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.210762 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.233419 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcd76a72-5869-4014-8bf7-ee57297feb85-config-data\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.233584 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dcd76a72-5869-4014-8bf7-ee57297feb85-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.233667 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dcd76a72-5869-4014-8bf7-ee57297feb85-public-tls-certs\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.233707 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcd76a72-5869-4014-8bf7-ee57297feb85-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.233824 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dcd76a72-5869-4014-8bf7-ee57297feb85-logs\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.233853 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dz58\" (UniqueName: \"kubernetes.io/projected/dcd76a72-5869-4014-8bf7-ee57297feb85-kube-api-access-2dz58\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.236865 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dcd76a72-5869-4014-8bf7-ee57297feb85-logs\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.239211 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.245020 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dcd76a72-5869-4014-8bf7-ee57297feb85-public-tls-certs\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.245125 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcd76a72-5869-4014-8bf7-ee57297feb85-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.245170 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dcd76a72-5869-4014-8bf7-ee57297feb85-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.245415 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcd76a72-5869-4014-8bf7-ee57297feb85-config-data\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.253164 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dz58\" (UniqueName: \"kubernetes.io/projected/dcd76a72-5869-4014-8bf7-ee57297feb85-kube-api-access-2dz58\") pod \"nova-api-0\" (UID: \"dcd76a72-5869-4014-8bf7-ee57297feb85\") " pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.336214 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a844a9-6868-43b6-b5a9-402e619f0d50-config-data\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.336415 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6tng\" (UniqueName: \"kubernetes.io/projected/f3a844a9-6868-43b6-b5a9-402e619f0d50-kube-api-access-z6tng\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.336474 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3a844a9-6868-43b6-b5a9-402e619f0d50-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.336761 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a844a9-6868-43b6-b5a9-402e619f0d50-logs\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.336845 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a844a9-6868-43b6-b5a9-402e619f0d50-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.345495 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.438669 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a844a9-6868-43b6-b5a9-402e619f0d50-logs\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.438736 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a844a9-6868-43b6-b5a9-402e619f0d50-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.438825 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a844a9-6868-43b6-b5a9-402e619f0d50-config-data\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.439300 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a844a9-6868-43b6-b5a9-402e619f0d50-logs\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.439462 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6tng\" (UniqueName: \"kubernetes.io/projected/f3a844a9-6868-43b6-b5a9-402e619f0d50-kube-api-access-z6tng\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.439507 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3a844a9-6868-43b6-b5a9-402e619f0d50-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.443855 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a844a9-6868-43b6-b5a9-402e619f0d50-config-data\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.445309 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a844a9-6868-43b6-b5a9-402e619f0d50-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.454266 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3a844a9-6868-43b6-b5a9-402e619f0d50-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.473240 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6tng\" (UniqueName: \"kubernetes.io/projected/f3a844a9-6868-43b6-b5a9-402e619f0d50-kube-api-access-z6tng\") pod \"nova-metadata-0\" (UID: \"f3a844a9-6868-43b6-b5a9-402e619f0d50\") " pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.665023 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.857375 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14d98749-4c8f-4ad4-abc4-3229198de8f7" path="/var/lib/kubelet/pods/14d98749-4c8f-4ad4-abc4-3229198de8f7/volumes" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.859115 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a7f9045-741f-4648-a9d5-8797c8415f7e" path="/var/lib/kubelet/pods/9a7f9045-741f-4648-a9d5-8797c8415f7e/volumes" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.860164 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c402ef80-d243-4a46-bd8f-9fa0036d771e" path="/var/lib/kubelet/pods/c402ef80-d243-4a46-bd8f-9fa0036d771e/volumes" Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.861618 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.861660 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2859495a-e026-400f-a304-05e25e961f2e","Type":"ContainerStarted","Data":"d0d4183c52f925037719ec6c846895be0337d955d643d43fbe47d57be81f9269"} Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.872980 4857 generic.go:334] "Generic (PLEG): container finished" podID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" containerID="71cfb60953babb168f9c7d7ab67e2585760786c88574fccd8a496bd665b95384" exitCode=0 Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.873062 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxcr6" event={"ID":"4e928e86-c08b-4b61-acbb-4a86fa4464a6","Type":"ContainerDied","Data":"71cfb60953babb168f9c7d7ab67e2585760786c88574fccd8a496bd665b95384"} Dec 01 21:58:17 crc kubenswrapper[4857]: W1201 21:58:17.884553 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddcd76a72_5869_4014_8bf7_ee57297feb85.slice/crio-3c3901dd7dff1abe51b9654a66ccb3ed80dc0835ce3c2b74083bc9da9f17b8e8 WatchSource:0}: Error finding container 3c3901dd7dff1abe51b9654a66ccb3ed80dc0835ce3c2b74083bc9da9f17b8e8: Status 404 returned error can't find the container with id 3c3901dd7dff1abe51b9654a66ccb3ed80dc0835ce3c2b74083bc9da9f17b8e8 Dec 01 21:58:17 crc kubenswrapper[4857]: I1201 21:58:17.904677 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.904650474 podStartE2EDuration="2.904650474s" podCreationTimestamp="2025-12-01 21:58:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:58:17.880643659 +0000 UTC m=+1416.370705976" watchObservedRunningTime="2025-12-01 21:58:17.904650474 +0000 UTC m=+1416.394712791" Dec 01 21:58:18 crc kubenswrapper[4857]: I1201 21:58:18.162718 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 21:58:18 crc kubenswrapper[4857]: I1201 21:58:18.889564 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a844a9-6868-43b6-b5a9-402e619f0d50","Type":"ContainerStarted","Data":"8495cd25732818824c989e644d47a957c124b57b26e44579c01db07d544815ee"} Dec 01 21:58:18 crc kubenswrapper[4857]: I1201 21:58:18.890419 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a844a9-6868-43b6-b5a9-402e619f0d50","Type":"ContainerStarted","Data":"5299aea0beecae846425d2436139da26efe2c3d14ea0b056361f66b809778093"} Dec 01 21:58:18 crc kubenswrapper[4857]: I1201 21:58:18.890498 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a844a9-6868-43b6-b5a9-402e619f0d50","Type":"ContainerStarted","Data":"b7c601cfc81a532a6455bbdad3055cbbea7f4e561c85a7fc3280e624a2f8cfbf"} Dec 01 21:58:18 crc kubenswrapper[4857]: I1201 21:58:18.892474 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dcd76a72-5869-4014-8bf7-ee57297feb85","Type":"ContainerStarted","Data":"ee12e2624301661bb4a0140aa4820e6581d8afd059b5b125b147d941636522ef"} Dec 01 21:58:18 crc kubenswrapper[4857]: I1201 21:58:18.892515 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dcd76a72-5869-4014-8bf7-ee57297feb85","Type":"ContainerStarted","Data":"762ffd062ed4f49eded6ad9c6d7a051d38b9b8f0c56fcae9ee281575cf234066"} Dec 01 21:58:18 crc kubenswrapper[4857]: I1201 21:58:18.892533 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dcd76a72-5869-4014-8bf7-ee57297feb85","Type":"ContainerStarted","Data":"3c3901dd7dff1abe51b9654a66ccb3ed80dc0835ce3c2b74083bc9da9f17b8e8"} Dec 01 21:58:18 crc kubenswrapper[4857]: I1201 21:58:18.905333 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxcr6" event={"ID":"4e928e86-c08b-4b61-acbb-4a86fa4464a6","Type":"ContainerStarted","Data":"06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971"} Dec 01 21:58:18 crc kubenswrapper[4857]: I1201 21:58:18.930335 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.930316344 podStartE2EDuration="1.930316344s" podCreationTimestamp="2025-12-01 21:58:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:58:18.925031653 +0000 UTC m=+1417.415093980" watchObservedRunningTime="2025-12-01 21:58:18.930316344 +0000 UTC m=+1417.420378661" Dec 01 21:58:18 crc kubenswrapper[4857]: I1201 21:58:18.997524 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zxcr6" podStartSLOduration=3.3616337769999998 podStartE2EDuration="5.997505539s" podCreationTimestamp="2025-12-01 21:58:13 +0000 UTC" firstStartedPulling="2025-12-01 21:58:15.81240781 +0000 UTC m=+1414.302470147" lastFinishedPulling="2025-12-01 21:58:18.448279582 +0000 UTC m=+1416.938341909" observedRunningTime="2025-12-01 21:58:18.973483284 +0000 UTC m=+1417.463545621" watchObservedRunningTime="2025-12-01 21:58:18.997505539 +0000 UTC m=+1417.487567856" Dec 01 21:58:19 crc kubenswrapper[4857]: I1201 21:58:19.003103 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.003084737 podStartE2EDuration="3.003084737s" podCreationTimestamp="2025-12-01 21:58:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:58:18.994295739 +0000 UTC m=+1417.484358066" watchObservedRunningTime="2025-12-01 21:58:19.003084737 +0000 UTC m=+1417.493147054" Dec 01 21:58:21 crc kubenswrapper[4857]: I1201 21:58:21.237518 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 21:58:22 crc kubenswrapper[4857]: I1201 21:58:22.665617 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 21:58:22 crc kubenswrapper[4857]: I1201 21:58:22.665943 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 21:58:24 crc kubenswrapper[4857]: I1201 21:58:24.059442 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:24 crc kubenswrapper[4857]: I1201 21:58:24.059603 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:25 crc kubenswrapper[4857]: I1201 21:58:25.111704 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zxcr6" podUID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" containerName="registry-server" probeResult="failure" output=< Dec 01 21:58:25 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Dec 01 21:58:25 crc kubenswrapper[4857]: > Dec 01 21:58:26 crc kubenswrapper[4857]: I1201 21:58:26.869069 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 21:58:26 crc kubenswrapper[4857]: I1201 21:58:26.925684 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 21:58:27 crc kubenswrapper[4857]: I1201 21:58:27.346203 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 21:58:27 crc kubenswrapper[4857]: I1201 21:58:27.346272 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 21:58:27 crc kubenswrapper[4857]: I1201 21:58:27.666341 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 21:58:27 crc kubenswrapper[4857]: I1201 21:58:27.666807 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 21:58:27 crc kubenswrapper[4857]: I1201 21:58:27.836587 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:58:27 crc kubenswrapper[4857]: I1201 21:58:27.836644 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:58:27 crc kubenswrapper[4857]: I1201 21:58:27.917708 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 21:58:27 crc kubenswrapper[4857]: I1201 21:58:27.960535 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 21:58:28 crc kubenswrapper[4857]: I1201 21:58:28.360392 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dcd76a72-5869-4014-8bf7-ee57297feb85" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.210:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 21:58:28 crc kubenswrapper[4857]: I1201 21:58:28.360809 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dcd76a72-5869-4014-8bf7-ee57297feb85" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.210:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 21:58:28 crc kubenswrapper[4857]: I1201 21:58:28.684218 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f3a844a9-6868-43b6-b5a9-402e619f0d50" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 21:58:28 crc kubenswrapper[4857]: I1201 21:58:28.684239 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f3a844a9-6868-43b6-b5a9-402e619f0d50" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 21:58:31 crc kubenswrapper[4857]: I1201 21:58:31.722786 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 21:58:31 crc kubenswrapper[4857]: I1201 21:58:31.723563 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="b2f9d1d0-36e7-44f0-9c03-3817bafe85ce" containerName="kube-state-metrics" containerID="cri-o://79728cdc0cb978317ab106a8a0a8d0a3b9c48a6a8f65af21ceb348ef5560e712" gracePeriod=30 Dec 01 21:58:31 crc kubenswrapper[4857]: I1201 21:58:31.928084 4857 generic.go:334] "Generic (PLEG): container finished" podID="b2f9d1d0-36e7-44f0-9c03-3817bafe85ce" containerID="79728cdc0cb978317ab106a8a0a8d0a3b9c48a6a8f65af21ceb348ef5560e712" exitCode=2 Dec 01 21:58:31 crc kubenswrapper[4857]: I1201 21:58:31.928137 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b2f9d1d0-36e7-44f0-9c03-3817bafe85ce","Type":"ContainerDied","Data":"79728cdc0cb978317ab106a8a0a8d0a3b9c48a6a8f65af21ceb348ef5560e712"} Dec 01 21:58:32 crc kubenswrapper[4857]: I1201 21:58:32.221266 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 21:58:32 crc kubenswrapper[4857]: I1201 21:58:32.333510 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrtbd\" (UniqueName: \"kubernetes.io/projected/b2f9d1d0-36e7-44f0-9c03-3817bafe85ce-kube-api-access-wrtbd\") pod \"b2f9d1d0-36e7-44f0-9c03-3817bafe85ce\" (UID: \"b2f9d1d0-36e7-44f0-9c03-3817bafe85ce\") " Dec 01 21:58:32 crc kubenswrapper[4857]: I1201 21:58:32.341110 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2f9d1d0-36e7-44f0-9c03-3817bafe85ce-kube-api-access-wrtbd" (OuterVolumeSpecName: "kube-api-access-wrtbd") pod "b2f9d1d0-36e7-44f0-9c03-3817bafe85ce" (UID: "b2f9d1d0-36e7-44f0-9c03-3817bafe85ce"). InnerVolumeSpecName "kube-api-access-wrtbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:58:32 crc kubenswrapper[4857]: I1201 21:58:32.437985 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrtbd\" (UniqueName: \"kubernetes.io/projected/b2f9d1d0-36e7-44f0-9c03-3817bafe85ce-kube-api-access-wrtbd\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:32 crc kubenswrapper[4857]: I1201 21:58:32.941625 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b2f9d1d0-36e7-44f0-9c03-3817bafe85ce","Type":"ContainerDied","Data":"95e7ecc583ad65a7a80d8d391a951aff950101e3eae5126402f1d5be3cdbc2ac"} Dec 01 21:58:32 crc kubenswrapper[4857]: I1201 21:58:32.941688 4857 scope.go:117] "RemoveContainer" containerID="79728cdc0cb978317ab106a8a0a8d0a3b9c48a6a8f65af21ceb348ef5560e712" Dec 01 21:58:32 crc kubenswrapper[4857]: I1201 21:58:32.943242 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 21:58:32 crc kubenswrapper[4857]: I1201 21:58:32.989166 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 21:58:32 crc kubenswrapper[4857]: I1201 21:58:32.997705 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.026272 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 21:58:33 crc kubenswrapper[4857]: E1201 21:58:33.027080 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2f9d1d0-36e7-44f0-9c03-3817bafe85ce" containerName="kube-state-metrics" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.027120 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2f9d1d0-36e7-44f0-9c03-3817bafe85ce" containerName="kube-state-metrics" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.027539 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2f9d1d0-36e7-44f0-9c03-3817bafe85ce" containerName="kube-state-metrics" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.028985 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.056277 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.056388 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.075213 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.155915 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfbeeaca-f2fe-41ed-bb4d-092cb86751d5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5\") " pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.156025 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/cfbeeaca-f2fe-41ed-bb4d-092cb86751d5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5\") " pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.156087 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vmc9\" (UniqueName: \"kubernetes.io/projected/cfbeeaca-f2fe-41ed-bb4d-092cb86751d5-kube-api-access-7vmc9\") pod \"kube-state-metrics-0\" (UID: \"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5\") " pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.156289 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbeeaca-f2fe-41ed-bb4d-092cb86751d5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5\") " pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.258289 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbeeaca-f2fe-41ed-bb4d-092cb86751d5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5\") " pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.258397 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfbeeaca-f2fe-41ed-bb4d-092cb86751d5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5\") " pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.258433 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/cfbeeaca-f2fe-41ed-bb4d-092cb86751d5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5\") " pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.258458 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vmc9\" (UniqueName: \"kubernetes.io/projected/cfbeeaca-f2fe-41ed-bb4d-092cb86751d5-kube-api-access-7vmc9\") pod \"kube-state-metrics-0\" (UID: \"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5\") " pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.263703 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/cfbeeaca-f2fe-41ed-bb4d-092cb86751d5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5\") " pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.264482 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbeeaca-f2fe-41ed-bb4d-092cb86751d5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5\") " pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.265837 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfbeeaca-f2fe-41ed-bb4d-092cb86751d5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5\") " pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.285224 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vmc9\" (UniqueName: \"kubernetes.io/projected/cfbeeaca-f2fe-41ed-bb4d-092cb86751d5-kube-api-access-7vmc9\") pod \"kube-state-metrics-0\" (UID: \"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5\") " pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.348269 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 21:58:33 crc kubenswrapper[4857]: I1201 21:58:33.849592 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2f9d1d0-36e7-44f0-9c03-3817bafe85ce" path="/var/lib/kubelet/pods/b2f9d1d0-36e7-44f0-9c03-3817bafe85ce/volumes" Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.019105 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 21:58:34 crc kubenswrapper[4857]: W1201 21:58:34.022395 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfbeeaca_f2fe_41ed_bb4d_092cb86751d5.slice/crio-5baebae418f3840113a930e2a99d0bb5c456ed8fc2c56e85101b56e922c4e36d WatchSource:0}: Error finding container 5baebae418f3840113a930e2a99d0bb5c456ed8fc2c56e85101b56e922c4e36d: Status 404 returned error can't find the container with id 5baebae418f3840113a930e2a99d0bb5c456ed8fc2c56e85101b56e922c4e36d Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.125081 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.181913 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.236663 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.237005 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="ceilometer-central-agent" containerID="cri-o://650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef" gracePeriod=30 Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.237510 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="proxy-httpd" containerID="cri-o://304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31" gracePeriod=30 Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.237571 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="sg-core" containerID="cri-o://3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba" gracePeriod=30 Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.237614 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="ceilometer-notification-agent" containerID="cri-o://d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c" gracePeriod=30 Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.377290 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zxcr6"] Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.967655 4857 generic.go:334] "Generic (PLEG): container finished" podID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerID="304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31" exitCode=0 Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.968230 4857 generic.go:334] "Generic (PLEG): container finished" podID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerID="3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba" exitCode=2 Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.968238 4857 generic.go:334] "Generic (PLEG): container finished" podID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerID="650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef" exitCode=0 Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.968285 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f","Type":"ContainerDied","Data":"304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31"} Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.968327 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f","Type":"ContainerDied","Data":"3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba"} Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.968338 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f","Type":"ContainerDied","Data":"650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef"} Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.971440 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5","Type":"ContainerStarted","Data":"a597a238f4c41e2d7449782110cd7d604c92eb49549605f99507c86d4d900a68"} Dec 01 21:58:34 crc kubenswrapper[4857]: I1201 21:58:34.971472 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cfbeeaca-f2fe-41ed-bb4d-092cb86751d5","Type":"ContainerStarted","Data":"5baebae418f3840113a930e2a99d0bb5c456ed8fc2c56e85101b56e922c4e36d"} Dec 01 21:58:35 crc kubenswrapper[4857]: I1201 21:58:35.000178 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.631192544 podStartE2EDuration="3.000157675s" podCreationTimestamp="2025-12-01 21:58:32 +0000 UTC" firstStartedPulling="2025-12-01 21:58:34.029450846 +0000 UTC m=+1432.519513163" lastFinishedPulling="2025-12-01 21:58:34.398415977 +0000 UTC m=+1432.888478294" observedRunningTime="2025-12-01 21:58:34.993019328 +0000 UTC m=+1433.483081645" watchObservedRunningTime="2025-12-01 21:58:35.000157675 +0000 UTC m=+1433.490220012" Dec 01 21:58:35 crc kubenswrapper[4857]: I1201 21:58:35.981772 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zxcr6" podUID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" containerName="registry-server" containerID="cri-o://06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971" gracePeriod=2 Dec 01 21:58:35 crc kubenswrapper[4857]: I1201 21:58:35.982175 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 21:58:36 crc kubenswrapper[4857]: I1201 21:58:36.439806 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:36 crc kubenswrapper[4857]: I1201 21:58:36.564333 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ncpj\" (UniqueName: \"kubernetes.io/projected/4e928e86-c08b-4b61-acbb-4a86fa4464a6-kube-api-access-6ncpj\") pod \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\" (UID: \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\") " Dec 01 21:58:36 crc kubenswrapper[4857]: I1201 21:58:36.564867 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e928e86-c08b-4b61-acbb-4a86fa4464a6-utilities\") pod \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\" (UID: \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\") " Dec 01 21:58:36 crc kubenswrapper[4857]: I1201 21:58:36.564895 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e928e86-c08b-4b61-acbb-4a86fa4464a6-catalog-content\") pod \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\" (UID: \"4e928e86-c08b-4b61-acbb-4a86fa4464a6\") " Dec 01 21:58:36 crc kubenswrapper[4857]: I1201 21:58:36.565825 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e928e86-c08b-4b61-acbb-4a86fa4464a6-utilities" (OuterVolumeSpecName: "utilities") pod "4e928e86-c08b-4b61-acbb-4a86fa4464a6" (UID: "4e928e86-c08b-4b61-acbb-4a86fa4464a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:58:36 crc kubenswrapper[4857]: I1201 21:58:36.570254 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e928e86-c08b-4b61-acbb-4a86fa4464a6-kube-api-access-6ncpj" (OuterVolumeSpecName: "kube-api-access-6ncpj") pod "4e928e86-c08b-4b61-acbb-4a86fa4464a6" (UID: "4e928e86-c08b-4b61-acbb-4a86fa4464a6"). InnerVolumeSpecName "kube-api-access-6ncpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:58:36 crc kubenswrapper[4857]: I1201 21:58:36.662700 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e928e86-c08b-4b61-acbb-4a86fa4464a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e928e86-c08b-4b61-acbb-4a86fa4464a6" (UID: "4e928e86-c08b-4b61-acbb-4a86fa4464a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:58:36 crc kubenswrapper[4857]: I1201 21:58:36.669083 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ncpj\" (UniqueName: \"kubernetes.io/projected/4e928e86-c08b-4b61-acbb-4a86fa4464a6-kube-api-access-6ncpj\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:36 crc kubenswrapper[4857]: I1201 21:58:36.669112 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e928e86-c08b-4b61-acbb-4a86fa4464a6-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:36 crc kubenswrapper[4857]: I1201 21:58:36.669125 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e928e86-c08b-4b61-acbb-4a86fa4464a6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.006514 4857 generic.go:334] "Generic (PLEG): container finished" podID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" containerID="06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971" exitCode=0 Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.007291 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxcr6" event={"ID":"4e928e86-c08b-4b61-acbb-4a86fa4464a6","Type":"ContainerDied","Data":"06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971"} Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.007353 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zxcr6" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.007378 4857 scope.go:117] "RemoveContainer" containerID="06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.007363 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxcr6" event={"ID":"4e928e86-c08b-4b61-acbb-4a86fa4464a6","Type":"ContainerDied","Data":"b8bf608264eca8094ed6d17c726019a788dec8e7f085b8318fe335aa2ff74760"} Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.050259 4857 scope.go:117] "RemoveContainer" containerID="71cfb60953babb168f9c7d7ab67e2585760786c88574fccd8a496bd665b95384" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.056767 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zxcr6"] Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.066196 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zxcr6"] Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.092030 4857 scope.go:117] "RemoveContainer" containerID="ca8babef73bb4515085c02382a7fa70fb49c68f7d9dacaac791f657746813158" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.117524 4857 scope.go:117] "RemoveContainer" containerID="06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971" Dec 01 21:58:37 crc kubenswrapper[4857]: E1201 21:58:37.117963 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971\": container with ID starting with 06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971 not found: ID does not exist" containerID="06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.118020 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971"} err="failed to get container status \"06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971\": rpc error: code = NotFound desc = could not find container \"06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971\": container with ID starting with 06fc1eda04979cfea12b39b0ec65969abaa5eebd70050674011f139122e77971 not found: ID does not exist" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.118073 4857 scope.go:117] "RemoveContainer" containerID="71cfb60953babb168f9c7d7ab67e2585760786c88574fccd8a496bd665b95384" Dec 01 21:58:37 crc kubenswrapper[4857]: E1201 21:58:37.118440 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71cfb60953babb168f9c7d7ab67e2585760786c88574fccd8a496bd665b95384\": container with ID starting with 71cfb60953babb168f9c7d7ab67e2585760786c88574fccd8a496bd665b95384 not found: ID does not exist" containerID="71cfb60953babb168f9c7d7ab67e2585760786c88574fccd8a496bd665b95384" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.118475 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71cfb60953babb168f9c7d7ab67e2585760786c88574fccd8a496bd665b95384"} err="failed to get container status \"71cfb60953babb168f9c7d7ab67e2585760786c88574fccd8a496bd665b95384\": rpc error: code = NotFound desc = could not find container \"71cfb60953babb168f9c7d7ab67e2585760786c88574fccd8a496bd665b95384\": container with ID starting with 71cfb60953babb168f9c7d7ab67e2585760786c88574fccd8a496bd665b95384 not found: ID does not exist" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.118499 4857 scope.go:117] "RemoveContainer" containerID="ca8babef73bb4515085c02382a7fa70fb49c68f7d9dacaac791f657746813158" Dec 01 21:58:37 crc kubenswrapper[4857]: E1201 21:58:37.118796 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca8babef73bb4515085c02382a7fa70fb49c68f7d9dacaac791f657746813158\": container with ID starting with ca8babef73bb4515085c02382a7fa70fb49c68f7d9dacaac791f657746813158 not found: ID does not exist" containerID="ca8babef73bb4515085c02382a7fa70fb49c68f7d9dacaac791f657746813158" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.118836 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca8babef73bb4515085c02382a7fa70fb49c68f7d9dacaac791f657746813158"} err="failed to get container status \"ca8babef73bb4515085c02382a7fa70fb49c68f7d9dacaac791f657746813158\": rpc error: code = NotFound desc = could not find container \"ca8babef73bb4515085c02382a7fa70fb49c68f7d9dacaac791f657746813158\": container with ID starting with ca8babef73bb4515085c02382a7fa70fb49c68f7d9dacaac791f657746813158 not found: ID does not exist" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.368100 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.369055 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.369517 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.385119 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.512454 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.590772 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-config-data\") pod \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.590857 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-log-httpd\") pod \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.590886 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-run-httpd\") pod \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.590986 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjz45\" (UniqueName: \"kubernetes.io/projected/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-kube-api-access-qjz45\") pod \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.591352 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-sg-core-conf-yaml\") pod \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.591402 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-combined-ca-bundle\") pod \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.591461 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-scripts\") pod \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\" (UID: \"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f\") " Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.591849 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" (UID: "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.592114 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" (UID: "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.592140 4857 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.604512 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-scripts" (OuterVolumeSpecName: "scripts") pod "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" (UID: "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.608739 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-kube-api-access-qjz45" (OuterVolumeSpecName: "kube-api-access-qjz45") pod "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" (UID: "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f"). InnerVolumeSpecName "kube-api-access-qjz45". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.637723 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" (UID: "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.675231 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.675533 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.684995 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.687263 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.690258 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" (UID: "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.695949 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.695978 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.695990 4857 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.696000 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjz45\" (UniqueName: \"kubernetes.io/projected/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-kube-api-access-qjz45\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.696010 4857 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.708949 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-config-data" (OuterVolumeSpecName: "config-data") pod "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" (UID: "cfbabfb2-c3fd-4242-8404-e32e6c3acf2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.798973 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:58:37 crc kubenswrapper[4857]: I1201 21:58:37.846836 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" path="/var/lib/kubelet/pods/4e928e86-c08b-4b61-acbb-4a86fa4464a6/volumes" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.021609 4857 generic.go:334] "Generic (PLEG): container finished" podID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerID="d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c" exitCode=0 Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.021719 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.021709 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f","Type":"ContainerDied","Data":"d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c"} Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.021889 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cfbabfb2-c3fd-4242-8404-e32e6c3acf2f","Type":"ContainerDied","Data":"5dabceb8e86ffdb2ab5ca8157e38fd8eb424d59a7b9bcc504d1670f29fc268c4"} Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.021933 4857 scope.go:117] "RemoveContainer" containerID="304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.024130 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.060230 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.062275 4857 scope.go:117] "RemoveContainer" containerID="3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.065404 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.096547 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.119714 4857 scope.go:117] "RemoveContainer" containerID="d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.148321 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:58:38 crc kubenswrapper[4857]: E1201 21:58:38.148959 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="ceilometer-notification-agent" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.148980 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="ceilometer-notification-agent" Dec 01 21:58:38 crc kubenswrapper[4857]: E1201 21:58:38.149004 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" containerName="extract-utilities" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.149015 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" containerName="extract-utilities" Dec 01 21:58:38 crc kubenswrapper[4857]: E1201 21:58:38.149027 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="ceilometer-central-agent" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.149057 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="ceilometer-central-agent" Dec 01 21:58:38 crc kubenswrapper[4857]: E1201 21:58:38.149070 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="proxy-httpd" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.149078 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="proxy-httpd" Dec 01 21:58:38 crc kubenswrapper[4857]: E1201 21:58:38.149091 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" containerName="registry-server" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.149096 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" containerName="registry-server" Dec 01 21:58:38 crc kubenswrapper[4857]: E1201 21:58:38.149118 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="sg-core" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.149126 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="sg-core" Dec 01 21:58:38 crc kubenswrapper[4857]: E1201 21:58:38.149176 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" containerName="extract-content" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.149184 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" containerName="extract-content" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.149462 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="proxy-httpd" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.149480 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="sg-core" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.149488 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="ceilometer-central-agent" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.149505 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e928e86-c08b-4b61-acbb-4a86fa4464a6" containerName="registry-server" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.149541 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" containerName="ceilometer-notification-agent" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.160375 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.165335 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.165566 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.165862 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.169839 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.187087 4857 scope.go:117] "RemoveContainer" containerID="650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.229383 4857 scope.go:117] "RemoveContainer" containerID="304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31" Dec 01 21:58:38 crc kubenswrapper[4857]: E1201 21:58:38.230117 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31\": container with ID starting with 304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31 not found: ID does not exist" containerID="304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.230251 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31"} err="failed to get container status \"304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31\": rpc error: code = NotFound desc = could not find container \"304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31\": container with ID starting with 304f907df9f5369f1d178591c5ef0de9078408fa275171556750c9c984695f31 not found: ID does not exist" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.230399 4857 scope.go:117] "RemoveContainer" containerID="3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba" Dec 01 21:58:38 crc kubenswrapper[4857]: E1201 21:58:38.231107 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba\": container with ID starting with 3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba not found: ID does not exist" containerID="3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.231171 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba"} err="failed to get container status \"3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba\": rpc error: code = NotFound desc = could not find container \"3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba\": container with ID starting with 3ed31c85937b1ee1bf2c5f04086b1ac2a01dd0e5bc11d4478ccad885d0ae8cba not found: ID does not exist" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.231212 4857 scope.go:117] "RemoveContainer" containerID="d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c" Dec 01 21:58:38 crc kubenswrapper[4857]: E1201 21:58:38.231736 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c\": container with ID starting with d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c not found: ID does not exist" containerID="d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.231765 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c"} err="failed to get container status \"d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c\": rpc error: code = NotFound desc = could not find container \"d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c\": container with ID starting with d2f31062258146440319f10aeb5aa9d906f6d68601a1062953f5487f3029c79c not found: ID does not exist" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.231783 4857 scope.go:117] "RemoveContainer" containerID="650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef" Dec 01 21:58:38 crc kubenswrapper[4857]: E1201 21:58:38.232163 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef\": container with ID starting with 650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef not found: ID does not exist" containerID="650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.232203 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef"} err="failed to get container status \"650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef\": rpc error: code = NotFound desc = could not find container \"650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef\": container with ID starting with 650ebc7c113f6da8a979806746ff55b1f3eaa5b7c310ea2f02b518513c424fef not found: ID does not exist" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.311839 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.311916 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed82b79a-baa2-42ed-a35c-935dfacc778f-run-httpd\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.311980 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.312025 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.312077 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74dws\" (UniqueName: \"kubernetes.io/projected/ed82b79a-baa2-42ed-a35c-935dfacc778f-kube-api-access-74dws\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.312168 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed82b79a-baa2-42ed-a35c-935dfacc778f-log-httpd\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.312202 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-config-data\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.312231 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-scripts\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.413793 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-config-data\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.414217 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-scripts\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.414385 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.414531 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed82b79a-baa2-42ed-a35c-935dfacc778f-run-httpd\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.414699 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.414840 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.414970 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74dws\" (UniqueName: \"kubernetes.io/projected/ed82b79a-baa2-42ed-a35c-935dfacc778f-kube-api-access-74dws\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.415588 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed82b79a-baa2-42ed-a35c-935dfacc778f-log-httpd\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.415721 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed82b79a-baa2-42ed-a35c-935dfacc778f-run-httpd\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.416395 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed82b79a-baa2-42ed-a35c-935dfacc778f-log-httpd\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.419259 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.420218 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.421898 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.423430 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-config-data\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.424799 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-scripts\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.439727 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74dws\" (UniqueName: \"kubernetes.io/projected/ed82b79a-baa2-42ed-a35c-935dfacc778f-kube-api-access-74dws\") pod \"ceilometer-0\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.497903 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 21:58:38 crc kubenswrapper[4857]: I1201 21:58:38.994705 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 21:58:39 crc kubenswrapper[4857]: W1201 21:58:39.001359 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded82b79a_baa2_42ed_a35c_935dfacc778f.slice/crio-3d447f8817e1f4b21e0e8c51cdcdcb49eb15cf8a0ad37df746ad030807b3b820 WatchSource:0}: Error finding container 3d447f8817e1f4b21e0e8c51cdcdcb49eb15cf8a0ad37df746ad030807b3b820: Status 404 returned error can't find the container with id 3d447f8817e1f4b21e0e8c51cdcdcb49eb15cf8a0ad37df746ad030807b3b820 Dec 01 21:58:39 crc kubenswrapper[4857]: I1201 21:58:39.035301 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed82b79a-baa2-42ed-a35c-935dfacc778f","Type":"ContainerStarted","Data":"3d447f8817e1f4b21e0e8c51cdcdcb49eb15cf8a0ad37df746ad030807b3b820"} Dec 01 21:58:39 crc kubenswrapper[4857]: I1201 21:58:39.868132 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfbabfb2-c3fd-4242-8404-e32e6c3acf2f" path="/var/lib/kubelet/pods/cfbabfb2-c3fd-4242-8404-e32e6c3acf2f/volumes" Dec 01 21:58:40 crc kubenswrapper[4857]: I1201 21:58:40.047198 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed82b79a-baa2-42ed-a35c-935dfacc778f","Type":"ContainerStarted","Data":"85615bd37db3ac0824557104e6025e64e6f8570c228d3a62f90bb002e61e6d86"} Dec 01 21:58:41 crc kubenswrapper[4857]: I1201 21:58:41.070103 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed82b79a-baa2-42ed-a35c-935dfacc778f","Type":"ContainerStarted","Data":"1953a1657ee82da7d20d764e11157274898e04446c6a3dd4976b6233c0e80149"} Dec 01 21:58:42 crc kubenswrapper[4857]: I1201 21:58:42.081891 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed82b79a-baa2-42ed-a35c-935dfacc778f","Type":"ContainerStarted","Data":"cb2192707cd1882c643688790bbc8111c00787d69d16d549232bd81f26c9e7d8"} Dec 01 21:58:43 crc kubenswrapper[4857]: I1201 21:58:43.364995 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 21:58:44 crc kubenswrapper[4857]: I1201 21:58:44.104127 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed82b79a-baa2-42ed-a35c-935dfacc778f","Type":"ContainerStarted","Data":"6e2ea6014b27950b384592d210206391045a8404ba8de5cda7ffeacd1e103116"} Dec 01 21:58:44 crc kubenswrapper[4857]: I1201 21:58:44.104707 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 21:58:44 crc kubenswrapper[4857]: I1201 21:58:44.139596 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.051785135 podStartE2EDuration="6.139516396s" podCreationTimestamp="2025-12-01 21:58:38 +0000 UTC" firstStartedPulling="2025-12-01 21:58:39.004365926 +0000 UTC m=+1437.494428233" lastFinishedPulling="2025-12-01 21:58:43.092097177 +0000 UTC m=+1441.582159494" observedRunningTime="2025-12-01 21:58:44.128934444 +0000 UTC m=+1442.618996751" watchObservedRunningTime="2025-12-01 21:58:44.139516396 +0000 UTC m=+1442.629578733" Dec 01 21:58:57 crc kubenswrapper[4857]: I1201 21:58:57.806593 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:58:57 crc kubenswrapper[4857]: I1201 21:58:57.807713 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:59:08 crc kubenswrapper[4857]: I1201 21:59:08.507936 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 21:59:18 crc kubenswrapper[4857]: I1201 21:59:18.677030 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 21:59:19 crc kubenswrapper[4857]: I1201 21:59:19.652704 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 21:59:23 crc kubenswrapper[4857]: I1201 21:59:23.751582 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="235e0415-1e77-480d-b3da-157ed0252d4f" containerName="rabbitmq" containerID="cri-o://61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce" gracePeriod=604795 Dec 01 21:59:24 crc kubenswrapper[4857]: I1201 21:59:24.264486 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="84a09f7d-b0a7-4055-93e1-83d11edb6467" containerName="rabbitmq" containerID="cri-o://55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb" gracePeriod=604796 Dec 01 21:59:27 crc kubenswrapper[4857]: I1201 21:59:27.805812 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 21:59:27 crc kubenswrapper[4857]: I1201 21:59:27.806210 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 21:59:27 crc kubenswrapper[4857]: I1201 21:59:27.806265 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 21:59:27 crc kubenswrapper[4857]: I1201 21:59:27.807242 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"022dacb43671109561ed218fe31f14692c5b447676b12b6fcdbccdc338c9650c"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 21:59:27 crc kubenswrapper[4857]: I1201 21:59:27.807324 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://022dacb43671109561ed218fe31f14692c5b447676b12b6fcdbccdc338c9650c" gracePeriod=600 Dec 01 21:59:27 crc kubenswrapper[4857]: I1201 21:59:27.974791 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-848dd"] Dec 01 21:59:27 crc kubenswrapper[4857]: I1201 21:59:27.979723 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:27 crc kubenswrapper[4857]: I1201 21:59:27.996467 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-848dd"] Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.027278 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qddxz\" (UniqueName: \"kubernetes.io/projected/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-kube-api-access-qddxz\") pod \"redhat-marketplace-848dd\" (UID: \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\") " pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.027593 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-catalog-content\") pod \"redhat-marketplace-848dd\" (UID: \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\") " pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.027761 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-utilities\") pod \"redhat-marketplace-848dd\" (UID: \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\") " pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.129576 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qddxz\" (UniqueName: \"kubernetes.io/projected/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-kube-api-access-qddxz\") pod \"redhat-marketplace-848dd\" (UID: \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\") " pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.129688 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-catalog-content\") pod \"redhat-marketplace-848dd\" (UID: \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\") " pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.129752 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-utilities\") pod \"redhat-marketplace-848dd\" (UID: \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\") " pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.130756 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-utilities\") pod \"redhat-marketplace-848dd\" (UID: \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\") " pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.130899 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-catalog-content\") pod \"redhat-marketplace-848dd\" (UID: \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\") " pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.165417 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qddxz\" (UniqueName: \"kubernetes.io/projected/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-kube-api-access-qddxz\") pod \"redhat-marketplace-848dd\" (UID: \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\") " pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.348369 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.774221 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="235e0415-1e77-480d-b3da-157ed0252d4f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 01 21:59:28 crc kubenswrapper[4857]: W1201 21:59:28.879900 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4f80b30_d0e9_4c6c_ab50_7136ae3a5005.slice/crio-15bf320b8ef06b4c698462a3b008702e3b2b5fbea0846ccf295e751b305b75ce WatchSource:0}: Error finding container 15bf320b8ef06b4c698462a3b008702e3b2b5fbea0846ccf295e751b305b75ce: Status 404 returned error can't find the container with id 15bf320b8ef06b4c698462a3b008702e3b2b5fbea0846ccf295e751b305b75ce Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.884978 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-848dd"] Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.908098 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="022dacb43671109561ed218fe31f14692c5b447676b12b6fcdbccdc338c9650c" exitCode=0 Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.908155 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"022dacb43671109561ed218fe31f14692c5b447676b12b6fcdbccdc338c9650c"} Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.908224 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35"} Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.908252 4857 scope.go:117] "RemoveContainer" containerID="ae0326e43bb280b472fd2b09c89e53e5301501138efae77ad81d099184e7a849" Dec 01 21:59:28 crc kubenswrapper[4857]: I1201 21:59:28.910120 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-848dd" event={"ID":"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005","Type":"ContainerStarted","Data":"15bf320b8ef06b4c698462a3b008702e3b2b5fbea0846ccf295e751b305b75ce"} Dec 01 21:59:29 crc kubenswrapper[4857]: I1201 21:59:29.680627 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="84a09f7d-b0a7-4055-93e1-83d11edb6467" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 01 21:59:29 crc kubenswrapper[4857]: I1201 21:59:29.934585 4857 generic.go:334] "Generic (PLEG): container finished" podID="c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" containerID="57358d4b5f3a265ff37d5cbbd40128e00574b42d3a563eb14a4716076e458dc7" exitCode=0 Dec 01 21:59:29 crc kubenswrapper[4857]: I1201 21:59:29.934627 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-848dd" event={"ID":"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005","Type":"ContainerDied","Data":"57358d4b5f3a265ff37d5cbbd40128e00574b42d3a563eb14a4716076e458dc7"} Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.390423 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.481933 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-plugins-conf\") pod \"235e0415-1e77-480d-b3da-157ed0252d4f\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.482009 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-erlang-cookie\") pod \"235e0415-1e77-480d-b3da-157ed0252d4f\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.482057 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"235e0415-1e77-480d-b3da-157ed0252d4f\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.482109 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-tls\") pod \"235e0415-1e77-480d-b3da-157ed0252d4f\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.482138 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-plugins\") pod \"235e0415-1e77-480d-b3da-157ed0252d4f\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.482179 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/235e0415-1e77-480d-b3da-157ed0252d4f-erlang-cookie-secret\") pod \"235e0415-1e77-480d-b3da-157ed0252d4f\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.482239 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/235e0415-1e77-480d-b3da-157ed0252d4f-pod-info\") pod \"235e0415-1e77-480d-b3da-157ed0252d4f\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.482266 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-server-conf\") pod \"235e0415-1e77-480d-b3da-157ed0252d4f\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.482331 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-confd\") pod \"235e0415-1e77-480d-b3da-157ed0252d4f\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.482425 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdqld\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-kube-api-access-qdqld\") pod \"235e0415-1e77-480d-b3da-157ed0252d4f\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.482542 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-config-data\") pod \"235e0415-1e77-480d-b3da-157ed0252d4f\" (UID: \"235e0415-1e77-480d-b3da-157ed0252d4f\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.482590 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "235e0415-1e77-480d-b3da-157ed0252d4f" (UID: "235e0415-1e77-480d-b3da-157ed0252d4f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.483639 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "235e0415-1e77-480d-b3da-157ed0252d4f" (UID: "235e0415-1e77-480d-b3da-157ed0252d4f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.483753 4857 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.490516 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "235e0415-1e77-480d-b3da-157ed0252d4f" (UID: "235e0415-1e77-480d-b3da-157ed0252d4f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.493898 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "235e0415-1e77-480d-b3da-157ed0252d4f" (UID: "235e0415-1e77-480d-b3da-157ed0252d4f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.500465 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/235e0415-1e77-480d-b3da-157ed0252d4f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "235e0415-1e77-480d-b3da-157ed0252d4f" (UID: "235e0415-1e77-480d-b3da-157ed0252d4f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.500651 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-kube-api-access-qdqld" (OuterVolumeSpecName: "kube-api-access-qdqld") pod "235e0415-1e77-480d-b3da-157ed0252d4f" (UID: "235e0415-1e77-480d-b3da-157ed0252d4f"). InnerVolumeSpecName "kube-api-access-qdqld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.501515 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "235e0415-1e77-480d-b3da-157ed0252d4f" (UID: "235e0415-1e77-480d-b3da-157ed0252d4f"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.502831 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/235e0415-1e77-480d-b3da-157ed0252d4f-pod-info" (OuterVolumeSpecName: "pod-info") pod "235e0415-1e77-480d-b3da-157ed0252d4f" (UID: "235e0415-1e77-480d-b3da-157ed0252d4f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.578146 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-config-data" (OuterVolumeSpecName: "config-data") pod "235e0415-1e77-480d-b3da-157ed0252d4f" (UID: "235e0415-1e77-480d-b3da-157ed0252d4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.586421 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdqld\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-kube-api-access-qdqld\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.586669 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.586733 4857 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.586819 4857 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.586909 4857 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.586993 4857 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.587078 4857 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/235e0415-1e77-480d-b3da-157ed0252d4f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.587166 4857 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/235e0415-1e77-480d-b3da-157ed0252d4f-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.617881 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-server-conf" (OuterVolumeSpecName: "server-conf") pod "235e0415-1e77-480d-b3da-157ed0252d4f" (UID: "235e0415-1e77-480d-b3da-157ed0252d4f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.631975 4857 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.689235 4857 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/235e0415-1e77-480d-b3da-157ed0252d4f-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.689278 4857 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.718215 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "235e0415-1e77-480d-b3da-157ed0252d4f" (UID: "235e0415-1e77-480d-b3da-157ed0252d4f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.794302 4857 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/235e0415-1e77-480d-b3da-157ed0252d4f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.911494 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.969209 4857 generic.go:334] "Generic (PLEG): container finished" podID="84a09f7d-b0a7-4055-93e1-83d11edb6467" containerID="55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb" exitCode=0 Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.969291 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"84a09f7d-b0a7-4055-93e1-83d11edb6467","Type":"ContainerDied","Data":"55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb"} Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.969300 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.969323 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"84a09f7d-b0a7-4055-93e1-83d11edb6467","Type":"ContainerDied","Data":"dfe281ab23930f2c45d9b50e610b8b139302b96aff16fe846f5a7a987ad525be"} Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.969341 4857 scope.go:117] "RemoveContainer" containerID="55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.994727 4857 generic.go:334] "Generic (PLEG): container finished" podID="235e0415-1e77-480d-b3da-157ed0252d4f" containerID="61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce" exitCode=0 Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.995152 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"235e0415-1e77-480d-b3da-157ed0252d4f","Type":"ContainerDied","Data":"61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce"} Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.995179 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"235e0415-1e77-480d-b3da-157ed0252d4f","Type":"ContainerDied","Data":"465ef4ddbbe189e930cce6323f28c5f8065568a5db7f8cb04fb69271dd658c94"} Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.995244 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.998341 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-config-data\") pod \"84a09f7d-b0a7-4055-93e1-83d11edb6467\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.998378 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-tls\") pod \"84a09f7d-b0a7-4055-93e1-83d11edb6467\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.998399 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-server-conf\") pod \"84a09f7d-b0a7-4055-93e1-83d11edb6467\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.998462 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-erlang-cookie\") pod \"84a09f7d-b0a7-4055-93e1-83d11edb6467\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.998488 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84a09f7d-b0a7-4055-93e1-83d11edb6467-pod-info\") pod \"84a09f7d-b0a7-4055-93e1-83d11edb6467\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.998595 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twjb6\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-kube-api-access-twjb6\") pod \"84a09f7d-b0a7-4055-93e1-83d11edb6467\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.998651 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"84a09f7d-b0a7-4055-93e1-83d11edb6467\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.998698 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-plugins-conf\") pod \"84a09f7d-b0a7-4055-93e1-83d11edb6467\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.998801 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-plugins\") pod \"84a09f7d-b0a7-4055-93e1-83d11edb6467\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.998818 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-confd\") pod \"84a09f7d-b0a7-4055-93e1-83d11edb6467\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.998842 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84a09f7d-b0a7-4055-93e1-83d11edb6467-erlang-cookie-secret\") pod \"84a09f7d-b0a7-4055-93e1-83d11edb6467\" (UID: \"84a09f7d-b0a7-4055-93e1-83d11edb6467\") " Dec 01 21:59:30 crc kubenswrapper[4857]: I1201 21:59:30.999886 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "84a09f7d-b0a7-4055-93e1-83d11edb6467" (UID: "84a09f7d-b0a7-4055-93e1-83d11edb6467"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.000254 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "84a09f7d-b0a7-4055-93e1-83d11edb6467" (UID: "84a09f7d-b0a7-4055-93e1-83d11edb6467"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.003976 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "84a09f7d-b0a7-4055-93e1-83d11edb6467" (UID: "84a09f7d-b0a7-4055-93e1-83d11edb6467"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.010315 4857 scope.go:117] "RemoveContainer" containerID="8e89121786dac33726d523e59f2c80de2165f73812834f8643c69bd32e7e2d1c" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.030067 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "84a09f7d-b0a7-4055-93e1-83d11edb6467" (UID: "84a09f7d-b0a7-4055-93e1-83d11edb6467"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.035424 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-kube-api-access-twjb6" (OuterVolumeSpecName: "kube-api-access-twjb6") pod "84a09f7d-b0a7-4055-93e1-83d11edb6467" (UID: "84a09f7d-b0a7-4055-93e1-83d11edb6467"). InnerVolumeSpecName "kube-api-access-twjb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.040085 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/84a09f7d-b0a7-4055-93e1-83d11edb6467-pod-info" (OuterVolumeSpecName: "pod-info") pod "84a09f7d-b0a7-4055-93e1-83d11edb6467" (UID: "84a09f7d-b0a7-4055-93e1-83d11edb6467"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.040388 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84a09f7d-b0a7-4055-93e1-83d11edb6467-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "84a09f7d-b0a7-4055-93e1-83d11edb6467" (UID: "84a09f7d-b0a7-4055-93e1-83d11edb6467"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.051467 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "84a09f7d-b0a7-4055-93e1-83d11edb6467" (UID: "84a09f7d-b0a7-4055-93e1-83d11edb6467"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.101357 4857 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.101389 4857 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.101399 4857 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84a09f7d-b0a7-4055-93e1-83d11edb6467-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.101407 4857 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.101417 4857 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.101426 4857 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84a09f7d-b0a7-4055-93e1-83d11edb6467-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.101434 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twjb6\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-kube-api-access-twjb6\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.101465 4857 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.171873 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-config-data" (OuterVolumeSpecName: "config-data") pod "84a09f7d-b0a7-4055-93e1-83d11edb6467" (UID: "84a09f7d-b0a7-4055-93e1-83d11edb6467"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.183379 4857 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.198357 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-server-conf" (OuterVolumeSpecName: "server-conf") pod "84a09f7d-b0a7-4055-93e1-83d11edb6467" (UID: "84a09f7d-b0a7-4055-93e1-83d11edb6467"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.203681 4857 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.203725 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.203743 4857 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84a09f7d-b0a7-4055-93e1-83d11edb6467-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.251094 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "84a09f7d-b0a7-4055-93e1-83d11edb6467" (UID: "84a09f7d-b0a7-4055-93e1-83d11edb6467"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.306027 4857 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84a09f7d-b0a7-4055-93e1-83d11edb6467-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.327393 4857 scope.go:117] "RemoveContainer" containerID="55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb" Dec 01 21:59:31 crc kubenswrapper[4857]: E1201 21:59:31.327911 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb\": container with ID starting with 55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb not found: ID does not exist" containerID="55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.327964 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb"} err="failed to get container status \"55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb\": rpc error: code = NotFound desc = could not find container \"55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb\": container with ID starting with 55b30900e3d7b24813a2200d81cd3d2134baa4e680278b5784ca5102018f67cb not found: ID does not exist" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.328000 4857 scope.go:117] "RemoveContainer" containerID="8e89121786dac33726d523e59f2c80de2165f73812834f8643c69bd32e7e2d1c" Dec 01 21:59:31 crc kubenswrapper[4857]: E1201 21:59:31.330294 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e89121786dac33726d523e59f2c80de2165f73812834f8643c69bd32e7e2d1c\": container with ID starting with 8e89121786dac33726d523e59f2c80de2165f73812834f8643c69bd32e7e2d1c not found: ID does not exist" containerID="8e89121786dac33726d523e59f2c80de2165f73812834f8643c69bd32e7e2d1c" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.330329 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e89121786dac33726d523e59f2c80de2165f73812834f8643c69bd32e7e2d1c"} err="failed to get container status \"8e89121786dac33726d523e59f2c80de2165f73812834f8643c69bd32e7e2d1c\": rpc error: code = NotFound desc = could not find container \"8e89121786dac33726d523e59f2c80de2165f73812834f8643c69bd32e7e2d1c\": container with ID starting with 8e89121786dac33726d523e59f2c80de2165f73812834f8643c69bd32e7e2d1c not found: ID does not exist" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.330352 4857 scope.go:117] "RemoveContainer" containerID="61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.408066 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.409372 4857 scope.go:117] "RemoveContainer" containerID="868e89c708562210a5a6e915928b232c0e7618e25b2d87a0152f36a0bcbc14ad" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.420585 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.431050 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.439084 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.452140 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 21:59:31 crc kubenswrapper[4857]: E1201 21:59:31.452888 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235e0415-1e77-480d-b3da-157ed0252d4f" containerName="rabbitmq" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.452913 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="235e0415-1e77-480d-b3da-157ed0252d4f" containerName="rabbitmq" Dec 01 21:59:31 crc kubenswrapper[4857]: E1201 21:59:31.452951 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84a09f7d-b0a7-4055-93e1-83d11edb6467" containerName="setup-container" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.452960 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="84a09f7d-b0a7-4055-93e1-83d11edb6467" containerName="setup-container" Dec 01 21:59:31 crc kubenswrapper[4857]: E1201 21:59:31.452995 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235e0415-1e77-480d-b3da-157ed0252d4f" containerName="setup-container" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.453003 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="235e0415-1e77-480d-b3da-157ed0252d4f" containerName="setup-container" Dec 01 21:59:31 crc kubenswrapper[4857]: E1201 21:59:31.453015 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84a09f7d-b0a7-4055-93e1-83d11edb6467" containerName="rabbitmq" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.453023 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="84a09f7d-b0a7-4055-93e1-83d11edb6467" containerName="rabbitmq" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.453318 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="235e0415-1e77-480d-b3da-157ed0252d4f" containerName="rabbitmq" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.453342 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="84a09f7d-b0a7-4055-93e1-83d11edb6467" containerName="rabbitmq" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.454880 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.456612 4857 scope.go:117] "RemoveContainer" containerID="61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce" Dec 01 21:59:31 crc kubenswrapper[4857]: E1201 21:59:31.458868 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce\": container with ID starting with 61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce not found: ID does not exist" containerID="61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.458948 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce"} err="failed to get container status \"61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce\": rpc error: code = NotFound desc = could not find container \"61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce\": container with ID starting with 61b0d14d5cdc9edfa2b5d07b519f3e5e93bd1962ab223b4f6b0087c92bfdf6ce not found: ID does not exist" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.459001 4857 scope.go:117] "RemoveContainer" containerID="868e89c708562210a5a6e915928b232c0e7618e25b2d87a0152f36a0bcbc14ad" Dec 01 21:59:31 crc kubenswrapper[4857]: E1201 21:59:31.461102 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"868e89c708562210a5a6e915928b232c0e7618e25b2d87a0152f36a0bcbc14ad\": container with ID starting with 868e89c708562210a5a6e915928b232c0e7618e25b2d87a0152f36a0bcbc14ad not found: ID does not exist" containerID="868e89c708562210a5a6e915928b232c0e7618e25b2d87a0152f36a0bcbc14ad" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.461140 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"868e89c708562210a5a6e915928b232c0e7618e25b2d87a0152f36a0bcbc14ad"} err="failed to get container status \"868e89c708562210a5a6e915928b232c0e7618e25b2d87a0152f36a0bcbc14ad\": rpc error: code = NotFound desc = could not find container \"868e89c708562210a5a6e915928b232c0e7618e25b2d87a0152f36a0bcbc14ad\": container with ID starting with 868e89c708562210a5a6e915928b232c0e7618e25b2d87a0152f36a0bcbc14ad not found: ID does not exist" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.461595 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.461845 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.461890 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.461993 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.462127 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zgfph" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.461996 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.462274 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.463147 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.478794 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.481295 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.485060 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.485177 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.495777 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.496071 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.496093 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.496243 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.496564 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-57hqb" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.510478 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a49ab0a8-09f6-4bfa-9276-c89606c0030e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.510523 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a49ab0a8-09f6-4bfa-9276-c89606c0030e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.510543 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a49ab0a8-09f6-4bfa-9276-c89606c0030e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.510570 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a49ab0a8-09f6-4bfa-9276-c89606c0030e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.510598 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a49ab0a8-09f6-4bfa-9276-c89606c0030e-config-data\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.510613 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a49ab0a8-09f6-4bfa-9276-c89606c0030e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.510636 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a49ab0a8-09f6-4bfa-9276-c89606c0030e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.510657 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a49ab0a8-09f6-4bfa-9276-c89606c0030e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.510684 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.510726 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9526\" (UniqueName: \"kubernetes.io/projected/a49ab0a8-09f6-4bfa-9276-c89606c0030e-kube-api-access-j9526\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.510769 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a49ab0a8-09f6-4bfa-9276-c89606c0030e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.521931 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.612726 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.612791 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.612827 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.612876 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.612903 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.612925 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9526\" (UniqueName: \"kubernetes.io/projected/a49ab0a8-09f6-4bfa-9276-c89606c0030e-kube-api-access-j9526\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613001 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613022 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a49ab0a8-09f6-4bfa-9276-c89606c0030e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613080 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613119 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613170 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a49ab0a8-09f6-4bfa-9276-c89606c0030e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613193 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nclqt\" (UniqueName: \"kubernetes.io/projected/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-kube-api-access-nclqt\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613223 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613255 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a49ab0a8-09f6-4bfa-9276-c89606c0030e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613278 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a49ab0a8-09f6-4bfa-9276-c89606c0030e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613309 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a49ab0a8-09f6-4bfa-9276-c89606c0030e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613331 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613358 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a49ab0a8-09f6-4bfa-9276-c89606c0030e-config-data\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613381 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a49ab0a8-09f6-4bfa-9276-c89606c0030e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613414 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a49ab0a8-09f6-4bfa-9276-c89606c0030e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613410 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613448 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a49ab0a8-09f6-4bfa-9276-c89606c0030e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.613472 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.614952 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a49ab0a8-09f6-4bfa-9276-c89606c0030e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.615281 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a49ab0a8-09f6-4bfa-9276-c89606c0030e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.615570 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a49ab0a8-09f6-4bfa-9276-c89606c0030e-config-data\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.616563 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a49ab0a8-09f6-4bfa-9276-c89606c0030e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.616788 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a49ab0a8-09f6-4bfa-9276-c89606c0030e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.618066 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a49ab0a8-09f6-4bfa-9276-c89606c0030e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.619836 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a49ab0a8-09f6-4bfa-9276-c89606c0030e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.621384 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a49ab0a8-09f6-4bfa-9276-c89606c0030e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.621946 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a49ab0a8-09f6-4bfa-9276-c89606c0030e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.637358 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9526\" (UniqueName: \"kubernetes.io/projected/a49ab0a8-09f6-4bfa-9276-c89606c0030e-kube-api-access-j9526\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.655921 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"a49ab0a8-09f6-4bfa-9276-c89606c0030e\") " pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.717912 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.717979 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.718021 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.718114 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.718136 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nclqt\" (UniqueName: \"kubernetes.io/projected/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-kube-api-access-nclqt\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.718205 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.718287 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.718309 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.718334 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.719003 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.719794 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.719822 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.720228 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.720397 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.720531 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.720669 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.720770 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.723014 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.724610 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.724964 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.725227 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.746350 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nclqt\" (UniqueName: \"kubernetes.io/projected/cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825-kube-api-access-nclqt\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.759139 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.785558 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.815716 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.859607 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="235e0415-1e77-480d-b3da-157ed0252d4f" path="/var/lib/kubelet/pods/235e0415-1e77-480d-b3da-157ed0252d4f/volumes" Dec 01 21:59:31 crc kubenswrapper[4857]: I1201 21:59:31.866938 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84a09f7d-b0a7-4055-93e1-83d11edb6467" path="/var/lib/kubelet/pods/84a09f7d-b0a7-4055-93e1-83d11edb6467/volumes" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.027730 4857 generic.go:334] "Generic (PLEG): container finished" podID="c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" containerID="556f36300e28d9c08081088aa295e208bbf8dae7e8d7b7cf0c2995c088ab2bd3" exitCode=0 Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.027801 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-848dd" event={"ID":"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005","Type":"ContainerDied","Data":"556f36300e28d9c08081088aa295e208bbf8dae7e8d7b7cf0c2995c088ab2bd3"} Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.326489 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 21:59:32 crc kubenswrapper[4857]: W1201 21:59:32.329943 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda49ab0a8_09f6_4bfa_9276_c89606c0030e.slice/crio-8bed97bd715dcffa648af48d3b351f9b31985829d46035a6383f1dc3be44816c WatchSource:0}: Error finding container 8bed97bd715dcffa648af48d3b351f9b31985829d46035a6383f1dc3be44816c: Status 404 returned error can't find the container with id 8bed97bd715dcffa648af48d3b351f9b31985829d46035a6383f1dc3be44816c Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.416068 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 21:59:32 crc kubenswrapper[4857]: W1201 21:59:32.419120 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbfa87b5_fc91_47a6_8b3a_08a0ba3c9825.slice/crio-bf7e0334d3c164a96b862f109cafbb9e86b65aa60701b4db1abfd70315c2fec7 WatchSource:0}: Error finding container bf7e0334d3c164a96b862f109cafbb9e86b65aa60701b4db1abfd70315c2fec7: Status 404 returned error can't find the container with id bf7e0334d3c164a96b862f109cafbb9e86b65aa60701b4db1abfd70315c2fec7 Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.730787 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-wq8c6"] Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.734095 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.736779 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.766178 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-wq8c6"] Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.845788 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.845852 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.845888 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-config\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.845941 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.845967 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57nrd\" (UniqueName: \"kubernetes.io/projected/84a4e964-d1fe-48f0-a76b-ea193fac9f44-kube-api-access-57nrd\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.846097 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.846158 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.948691 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.949638 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.949820 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.949923 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.950084 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-config\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.950434 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.950530 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57nrd\" (UniqueName: \"kubernetes.io/projected/84a4e964-d1fe-48f0-a76b-ea193fac9f44-kube-api-access-57nrd\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.951269 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.951466 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.951640 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.951906 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-config\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.952430 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.952741 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:32 crc kubenswrapper[4857]: I1201 21:59:32.977184 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57nrd\" (UniqueName: \"kubernetes.io/projected/84a4e964-d1fe-48f0-a76b-ea193fac9f44-kube-api-access-57nrd\") pod \"dnsmasq-dns-7d84b4d45c-wq8c6\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:33 crc kubenswrapper[4857]: I1201 21:59:33.067762 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825","Type":"ContainerStarted","Data":"bf7e0334d3c164a96b862f109cafbb9e86b65aa60701b4db1abfd70315c2fec7"} Dec 01 21:59:33 crc kubenswrapper[4857]: I1201 21:59:33.075160 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-848dd" event={"ID":"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005","Type":"ContainerStarted","Data":"81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7"} Dec 01 21:59:33 crc kubenswrapper[4857]: I1201 21:59:33.075397 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:33 crc kubenswrapper[4857]: I1201 21:59:33.084484 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a49ab0a8-09f6-4bfa-9276-c89606c0030e","Type":"ContainerStarted","Data":"8bed97bd715dcffa648af48d3b351f9b31985829d46035a6383f1dc3be44816c"} Dec 01 21:59:33 crc kubenswrapper[4857]: I1201 21:59:33.622640 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-848dd" podStartSLOduration=3.992624497 podStartE2EDuration="6.622605996s" podCreationTimestamp="2025-12-01 21:59:27 +0000 UTC" firstStartedPulling="2025-12-01 21:59:29.937125844 +0000 UTC m=+1488.427188201" lastFinishedPulling="2025-12-01 21:59:32.567107383 +0000 UTC m=+1491.057169700" observedRunningTime="2025-12-01 21:59:33.10709754 +0000 UTC m=+1491.597159917" watchObservedRunningTime="2025-12-01 21:59:33.622605996 +0000 UTC m=+1492.112668313" Dec 01 21:59:33 crc kubenswrapper[4857]: I1201 21:59:33.631335 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-wq8c6"] Dec 01 21:59:34 crc kubenswrapper[4857]: I1201 21:59:34.101843 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" event={"ID":"84a4e964-d1fe-48f0-a76b-ea193fac9f44","Type":"ContainerStarted","Data":"ff6ed161048bd82d20bdc1e74d901cfe80796e80e9fcb4de48bc9b14c522ab6a"} Dec 01 21:59:35 crc kubenswrapper[4857]: I1201 21:59:35.115744 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825","Type":"ContainerStarted","Data":"78c166ce31b5bcf61a3dc702f2f0983511e762ceea37f6d54b3a49288de19622"} Dec 01 21:59:35 crc kubenswrapper[4857]: I1201 21:59:35.121391 4857 generic.go:334] "Generic (PLEG): container finished" podID="84a4e964-d1fe-48f0-a76b-ea193fac9f44" containerID="dd26e7f56729f2b810b5b4d6c4062cbe94533fb3a1033c5c77df2624194787f1" exitCode=0 Dec 01 21:59:35 crc kubenswrapper[4857]: I1201 21:59:35.121524 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" event={"ID":"84a4e964-d1fe-48f0-a76b-ea193fac9f44","Type":"ContainerDied","Data":"dd26e7f56729f2b810b5b4d6c4062cbe94533fb3a1033c5c77df2624194787f1"} Dec 01 21:59:35 crc kubenswrapper[4857]: I1201 21:59:35.129353 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a49ab0a8-09f6-4bfa-9276-c89606c0030e","Type":"ContainerStarted","Data":"3aee4b029f199f947d36fe646d64d6b2344b0e5ce2abc80e1e3a8e474346b46f"} Dec 01 21:59:36 crc kubenswrapper[4857]: I1201 21:59:36.144895 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" event={"ID":"84a4e964-d1fe-48f0-a76b-ea193fac9f44","Type":"ContainerStarted","Data":"7e4657a3c7e681a255f82c8795900d576db2a774ab704fc3812ad8867925b769"} Dec 01 21:59:36 crc kubenswrapper[4857]: I1201 21:59:36.190486 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" podStartSLOduration=4.190459539 podStartE2EDuration="4.190459539s" podCreationTimestamp="2025-12-01 21:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:59:36.178370347 +0000 UTC m=+1494.668432684" watchObservedRunningTime="2025-12-01 21:59:36.190459539 +0000 UTC m=+1494.680521856" Dec 01 21:59:37 crc kubenswrapper[4857]: I1201 21:59:37.170473 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:38 crc kubenswrapper[4857]: I1201 21:59:38.349280 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:38 crc kubenswrapper[4857]: I1201 21:59:38.350627 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:38 crc kubenswrapper[4857]: I1201 21:59:38.400478 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:39 crc kubenswrapper[4857]: I1201 21:59:39.241250 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:39 crc kubenswrapper[4857]: I1201 21:59:39.299384 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-848dd"] Dec 01 21:59:41 crc kubenswrapper[4857]: I1201 21:59:41.213315 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-848dd" podUID="c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" containerName="registry-server" containerID="cri-o://81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7" gracePeriod=2 Dec 01 21:59:41 crc kubenswrapper[4857]: I1201 21:59:41.700170 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:41 crc kubenswrapper[4857]: I1201 21:59:41.754137 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-catalog-content\") pod \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\" (UID: \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\") " Dec 01 21:59:41 crc kubenswrapper[4857]: I1201 21:59:41.754377 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-utilities\") pod \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\" (UID: \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\") " Dec 01 21:59:41 crc kubenswrapper[4857]: I1201 21:59:41.754423 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qddxz\" (UniqueName: \"kubernetes.io/projected/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-kube-api-access-qddxz\") pod \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\" (UID: \"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005\") " Dec 01 21:59:41 crc kubenswrapper[4857]: I1201 21:59:41.756125 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-utilities" (OuterVolumeSpecName: "utilities") pod "c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" (UID: "c4f80b30-d0e9-4c6c-ab50-7136ae3a5005"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:59:41 crc kubenswrapper[4857]: I1201 21:59:41.761436 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-kube-api-access-qddxz" (OuterVolumeSpecName: "kube-api-access-qddxz") pod "c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" (UID: "c4f80b30-d0e9-4c6c-ab50-7136ae3a5005"). InnerVolumeSpecName "kube-api-access-qddxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:59:41 crc kubenswrapper[4857]: I1201 21:59:41.783388 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" (UID: "c4f80b30-d0e9-4c6c-ab50-7136ae3a5005"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 21:59:41 crc kubenswrapper[4857]: I1201 21:59:41.856841 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:41 crc kubenswrapper[4857]: I1201 21:59:41.857226 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:41 crc kubenswrapper[4857]: I1201 21:59:41.857241 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qddxz\" (UniqueName: \"kubernetes.io/projected/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005-kube-api-access-qddxz\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.232366 4857 generic.go:334] "Generic (PLEG): container finished" podID="c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" containerID="81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7" exitCode=0 Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.232411 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-848dd" event={"ID":"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005","Type":"ContainerDied","Data":"81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7"} Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.232455 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-848dd" event={"ID":"c4f80b30-d0e9-4c6c-ab50-7136ae3a5005","Type":"ContainerDied","Data":"15bf320b8ef06b4c698462a3b008702e3b2b5fbea0846ccf295e751b305b75ce"} Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.232479 4857 scope.go:117] "RemoveContainer" containerID="81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7" Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.232474 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-848dd" Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.266960 4857 scope.go:117] "RemoveContainer" containerID="556f36300e28d9c08081088aa295e208bbf8dae7e8d7b7cf0c2995c088ab2bd3" Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.277966 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-848dd"] Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.294334 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-848dd"] Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.302022 4857 scope.go:117] "RemoveContainer" containerID="57358d4b5f3a265ff37d5cbbd40128e00574b42d3a563eb14a4716076e458dc7" Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.340449 4857 scope.go:117] "RemoveContainer" containerID="81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7" Dec 01 21:59:42 crc kubenswrapper[4857]: E1201 21:59:42.340878 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7\": container with ID starting with 81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7 not found: ID does not exist" containerID="81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7" Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.340914 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7"} err="failed to get container status \"81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7\": rpc error: code = NotFound desc = could not find container \"81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7\": container with ID starting with 81186b38aeb72263b769445a226e478ed9a75328d2f77d02bb43dfafee656ea7 not found: ID does not exist" Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.340940 4857 scope.go:117] "RemoveContainer" containerID="556f36300e28d9c08081088aa295e208bbf8dae7e8d7b7cf0c2995c088ab2bd3" Dec 01 21:59:42 crc kubenswrapper[4857]: E1201 21:59:42.341284 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"556f36300e28d9c08081088aa295e208bbf8dae7e8d7b7cf0c2995c088ab2bd3\": container with ID starting with 556f36300e28d9c08081088aa295e208bbf8dae7e8d7b7cf0c2995c088ab2bd3 not found: ID does not exist" containerID="556f36300e28d9c08081088aa295e208bbf8dae7e8d7b7cf0c2995c088ab2bd3" Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.341315 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"556f36300e28d9c08081088aa295e208bbf8dae7e8d7b7cf0c2995c088ab2bd3"} err="failed to get container status \"556f36300e28d9c08081088aa295e208bbf8dae7e8d7b7cf0c2995c088ab2bd3\": rpc error: code = NotFound desc = could not find container \"556f36300e28d9c08081088aa295e208bbf8dae7e8d7b7cf0c2995c088ab2bd3\": container with ID starting with 556f36300e28d9c08081088aa295e208bbf8dae7e8d7b7cf0c2995c088ab2bd3 not found: ID does not exist" Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.341333 4857 scope.go:117] "RemoveContainer" containerID="57358d4b5f3a265ff37d5cbbd40128e00574b42d3a563eb14a4716076e458dc7" Dec 01 21:59:42 crc kubenswrapper[4857]: E1201 21:59:42.341555 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57358d4b5f3a265ff37d5cbbd40128e00574b42d3a563eb14a4716076e458dc7\": container with ID starting with 57358d4b5f3a265ff37d5cbbd40128e00574b42d3a563eb14a4716076e458dc7 not found: ID does not exist" containerID="57358d4b5f3a265ff37d5cbbd40128e00574b42d3a563eb14a4716076e458dc7" Dec 01 21:59:42 crc kubenswrapper[4857]: I1201 21:59:42.341585 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57358d4b5f3a265ff37d5cbbd40128e00574b42d3a563eb14a4716076e458dc7"} err="failed to get container status \"57358d4b5f3a265ff37d5cbbd40128e00574b42d3a563eb14a4716076e458dc7\": rpc error: code = NotFound desc = could not find container \"57358d4b5f3a265ff37d5cbbd40128e00574b42d3a563eb14a4716076e458dc7\": container with ID starting with 57358d4b5f3a265ff37d5cbbd40128e00574b42d3a563eb14a4716076e458dc7 not found: ID does not exist" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.078211 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.146095 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx"] Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.146344 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" podUID="3325d8c6-37b4-40e4-819a-8dff28a9c6c9" containerName="dnsmasq-dns" containerID="cri-o://c5201045f96c5f52aacbe88de9e20041a3180be209abf5cf0d9f83ee7d48d100" gracePeriod=10 Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.359637 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-5mp7g"] Dec 01 21:59:43 crc kubenswrapper[4857]: E1201 21:59:43.360680 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" containerName="registry-server" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.360768 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" containerName="registry-server" Dec 01 21:59:43 crc kubenswrapper[4857]: E1201 21:59:43.360843 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" containerName="extract-utilities" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.360918 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" containerName="extract-utilities" Dec 01 21:59:43 crc kubenswrapper[4857]: E1201 21:59:43.360998 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" containerName="extract-content" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.361102 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" containerName="extract-content" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.361480 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" containerName="registry-server" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.371166 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.381618 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-5mp7g"] Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.507246 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.507376 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-config\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.507416 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.507454 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.507526 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfx9f\" (UniqueName: \"kubernetes.io/projected/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-kube-api-access-cfx9f\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.507603 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.507646 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.609302 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfx9f\" (UniqueName: \"kubernetes.io/projected/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-kube-api-access-cfx9f\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.609420 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.609468 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.609498 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.609561 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-config\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.609589 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.609650 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.611063 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.612317 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.616222 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.616821 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-config\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.616938 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.617626 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.654936 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfx9f\" (UniqueName: \"kubernetes.io/projected/72ba8edd-5f69-4c93-9954-aaa52e18c2fa-kube-api-access-cfx9f\") pod \"dnsmasq-dns-6f6df4f56c-5mp7g\" (UID: \"72ba8edd-5f69-4c93-9954-aaa52e18c2fa\") " pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.696499 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:43 crc kubenswrapper[4857]: I1201 21:59:43.857784 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4f80b30-d0e9-4c6c-ab50-7136ae3a5005" path="/var/lib/kubelet/pods/c4f80b30-d0e9-4c6c-ab50-7136ae3a5005/volumes" Dec 01 21:59:44 crc kubenswrapper[4857]: I1201 21:59:44.277669 4857 generic.go:334] "Generic (PLEG): container finished" podID="3325d8c6-37b4-40e4-819a-8dff28a9c6c9" containerID="c5201045f96c5f52aacbe88de9e20041a3180be209abf5cf0d9f83ee7d48d100" exitCode=0 Dec 01 21:59:44 crc kubenswrapper[4857]: I1201 21:59:44.277713 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" event={"ID":"3325d8c6-37b4-40e4-819a-8dff28a9c6c9","Type":"ContainerDied","Data":"c5201045f96c5f52aacbe88de9e20041a3180be209abf5cf0d9f83ee7d48d100"} Dec 01 21:59:44 crc kubenswrapper[4857]: W1201 21:59:44.666115 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72ba8edd_5f69_4c93_9954_aaa52e18c2fa.slice/crio-d9f0eb699e25ca821dc3004dbbe875636c6155d60362550e40137b5994eb46bf WatchSource:0}: Error finding container d9f0eb699e25ca821dc3004dbbe875636c6155d60362550e40137b5994eb46bf: Status 404 returned error can't find the container with id d9f0eb699e25ca821dc3004dbbe875636c6155d60362550e40137b5994eb46bf Dec 01 21:59:44 crc kubenswrapper[4857]: I1201 21:59:44.669829 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-5mp7g"] Dec 01 21:59:44 crc kubenswrapper[4857]: I1201 21:59:44.882088 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:59:44 crc kubenswrapper[4857]: I1201 21:59:44.948614 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-dns-svc\") pod \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " Dec 01 21:59:44 crc kubenswrapper[4857]: I1201 21:59:44.948710 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-ovsdbserver-nb\") pod \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " Dec 01 21:59:44 crc kubenswrapper[4857]: I1201 21:59:44.948827 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6brv5\" (UniqueName: \"kubernetes.io/projected/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-kube-api-access-6brv5\") pod \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " Dec 01 21:59:44 crc kubenswrapper[4857]: I1201 21:59:44.948909 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-config\") pod \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " Dec 01 21:59:44 crc kubenswrapper[4857]: I1201 21:59:44.949432 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-dns-swift-storage-0\") pod \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " Dec 01 21:59:44 crc kubenswrapper[4857]: I1201 21:59:44.949663 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-ovsdbserver-sb\") pod \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\" (UID: \"3325d8c6-37b4-40e4-819a-8dff28a9c6c9\") " Dec 01 21:59:44 crc kubenswrapper[4857]: I1201 21:59:44.978589 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-kube-api-access-6brv5" (OuterVolumeSpecName: "kube-api-access-6brv5") pod "3325d8c6-37b4-40e4-819a-8dff28a9c6c9" (UID: "3325d8c6-37b4-40e4-819a-8dff28a9c6c9"). InnerVolumeSpecName "kube-api-access-6brv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.029777 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3325d8c6-37b4-40e4-819a-8dff28a9c6c9" (UID: "3325d8c6-37b4-40e4-819a-8dff28a9c6c9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.032787 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3325d8c6-37b4-40e4-819a-8dff28a9c6c9" (UID: "3325d8c6-37b4-40e4-819a-8dff28a9c6c9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.035372 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3325d8c6-37b4-40e4-819a-8dff28a9c6c9" (UID: "3325d8c6-37b4-40e4-819a-8dff28a9c6c9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.039594 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-config" (OuterVolumeSpecName: "config") pod "3325d8c6-37b4-40e4-819a-8dff28a9c6c9" (UID: "3325d8c6-37b4-40e4-819a-8dff28a9c6c9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.051765 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3325d8c6-37b4-40e4-819a-8dff28a9c6c9" (UID: "3325d8c6-37b4-40e4-819a-8dff28a9c6c9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.058262 4857 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.058295 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.058306 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.058314 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.058323 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6brv5\" (UniqueName: \"kubernetes.io/projected/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-kube-api-access-6brv5\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.058332 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3325d8c6-37b4-40e4-819a-8dff28a9c6c9-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.294931 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" event={"ID":"3325d8c6-37b4-40e4-819a-8dff28a9c6c9","Type":"ContainerDied","Data":"bf39ecdfcb76014e08a7fa23067c20b156106b06d832f5e0b1b6b59ea4b23395"} Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.294968 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.295094 4857 scope.go:117] "RemoveContainer" containerID="c5201045f96c5f52aacbe88de9e20041a3180be209abf5cf0d9f83ee7d48d100" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.297428 4857 generic.go:334] "Generic (PLEG): container finished" podID="72ba8edd-5f69-4c93-9954-aaa52e18c2fa" containerID="f3619ba038f861a00fdbcdbab7ded0daba961cadbb64143680f610a762fac3fe" exitCode=0 Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.297490 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" event={"ID":"72ba8edd-5f69-4c93-9954-aaa52e18c2fa","Type":"ContainerDied","Data":"f3619ba038f861a00fdbcdbab7ded0daba961cadbb64143680f610a762fac3fe"} Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.297529 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" event={"ID":"72ba8edd-5f69-4c93-9954-aaa52e18c2fa","Type":"ContainerStarted","Data":"d9f0eb699e25ca821dc3004dbbe875636c6155d60362550e40137b5994eb46bf"} Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.321950 4857 scope.go:117] "RemoveContainer" containerID="6555feb690a2957e02e1bec2d9a2c09ba4d6fea69f4339c749c18aaab07d7de5" Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.353519 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx"] Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.363566 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-cc9tx"] Dec 01 21:59:45 crc kubenswrapper[4857]: I1201 21:59:45.865415 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3325d8c6-37b4-40e4-819a-8dff28a9c6c9" path="/var/lib/kubelet/pods/3325d8c6-37b4-40e4-819a-8dff28a9c6c9/volumes" Dec 01 21:59:46 crc kubenswrapper[4857]: I1201 21:59:46.311849 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" event={"ID":"72ba8edd-5f69-4c93-9954-aaa52e18c2fa","Type":"ContainerStarted","Data":"1baa04c22a44557739815cf2303f5645034d62ede62f9d5a5161724ffa5c166f"} Dec 01 21:59:46 crc kubenswrapper[4857]: I1201 21:59:46.311945 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:46 crc kubenswrapper[4857]: I1201 21:59:46.336232 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" podStartSLOduration=3.3362039660000002 podStartE2EDuration="3.336203966s" podCreationTimestamp="2025-12-01 21:59:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 21:59:46.333443292 +0000 UTC m=+1504.823505679" watchObservedRunningTime="2025-12-01 21:59:46.336203966 +0000 UTC m=+1504.826266323" Dec 01 21:59:51 crc kubenswrapper[4857]: E1201 21:59:51.367228 4857 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4f80b30_d0e9_4c6c_ab50_7136ae3a5005.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4f80b30_d0e9_4c6c_ab50_7136ae3a5005.slice/crio-15bf320b8ef06b4c698462a3b008702e3b2b5fbea0846ccf295e751b305b75ce\": RecentStats: unable to find data in memory cache]" Dec 01 21:59:53 crc kubenswrapper[4857]: I1201 21:59:53.698455 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f6df4f56c-5mp7g" Dec 01 21:59:53 crc kubenswrapper[4857]: I1201 21:59:53.783004 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-wq8c6"] Dec 01 21:59:53 crc kubenswrapper[4857]: I1201 21:59:53.784933 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" podUID="84a4e964-d1fe-48f0-a76b-ea193fac9f44" containerName="dnsmasq-dns" containerID="cri-o://7e4657a3c7e681a255f82c8795900d576db2a774ab704fc3812ad8867925b769" gracePeriod=10 Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.408760 4857 generic.go:334] "Generic (PLEG): container finished" podID="84a4e964-d1fe-48f0-a76b-ea193fac9f44" containerID="7e4657a3c7e681a255f82c8795900d576db2a774ab704fc3812ad8867925b769" exitCode=0 Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.409497 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" event={"ID":"84a4e964-d1fe-48f0-a76b-ea193fac9f44","Type":"ContainerDied","Data":"7e4657a3c7e681a255f82c8795900d576db2a774ab704fc3812ad8867925b769"} Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.409598 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" event={"ID":"84a4e964-d1fe-48f0-a76b-ea193fac9f44","Type":"ContainerDied","Data":"ff6ed161048bd82d20bdc1e74d901cfe80796e80e9fcb4de48bc9b14c522ab6a"} Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.409610 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff6ed161048bd82d20bdc1e74d901cfe80796e80e9fcb4de48bc9b14c522ab6a" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.448912 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.644027 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-openstack-edpm-ipam\") pod \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.644190 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-dns-svc\") pod \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.644295 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-config\") pod \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.644351 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-ovsdbserver-sb\") pod \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.644474 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-dns-swift-storage-0\") pod \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.644546 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-ovsdbserver-nb\") pod \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.644577 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57nrd\" (UniqueName: \"kubernetes.io/projected/84a4e964-d1fe-48f0-a76b-ea193fac9f44-kube-api-access-57nrd\") pod \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\" (UID: \"84a4e964-d1fe-48f0-a76b-ea193fac9f44\") " Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.652133 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84a4e964-d1fe-48f0-a76b-ea193fac9f44-kube-api-access-57nrd" (OuterVolumeSpecName: "kube-api-access-57nrd") pod "84a4e964-d1fe-48f0-a76b-ea193fac9f44" (UID: "84a4e964-d1fe-48f0-a76b-ea193fac9f44"). InnerVolumeSpecName "kube-api-access-57nrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.703371 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "84a4e964-d1fe-48f0-a76b-ea193fac9f44" (UID: "84a4e964-d1fe-48f0-a76b-ea193fac9f44"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.704294 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "84a4e964-d1fe-48f0-a76b-ea193fac9f44" (UID: "84a4e964-d1fe-48f0-a76b-ea193fac9f44"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.732700 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "84a4e964-d1fe-48f0-a76b-ea193fac9f44" (UID: "84a4e964-d1fe-48f0-a76b-ea193fac9f44"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.747931 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.748065 4857 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.748121 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57nrd\" (UniqueName: \"kubernetes.io/projected/84a4e964-d1fe-48f0-a76b-ea193fac9f44-kube-api-access-57nrd\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.748173 4857 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.762507 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "84a4e964-d1fe-48f0-a76b-ea193fac9f44" (UID: "84a4e964-d1fe-48f0-a76b-ea193fac9f44"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.782743 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "84a4e964-d1fe-48f0-a76b-ea193fac9f44" (UID: "84a4e964-d1fe-48f0-a76b-ea193fac9f44"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.795691 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-config" (OuterVolumeSpecName: "config") pod "84a4e964-d1fe-48f0-a76b-ea193fac9f44" (UID: "84a4e964-d1fe-48f0-a76b-ea193fac9f44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.850437 4857 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.850469 4857 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:54 crc kubenswrapper[4857]: I1201 21:59:54.850478 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a4e964-d1fe-48f0-a76b-ea193fac9f44-config\") on node \"crc\" DevicePath \"\"" Dec 01 21:59:55 crc kubenswrapper[4857]: I1201 21:59:55.433777 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-wq8c6" Dec 01 21:59:55 crc kubenswrapper[4857]: I1201 21:59:55.495633 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-wq8c6"] Dec 01 21:59:55 crc kubenswrapper[4857]: I1201 21:59:55.512337 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-wq8c6"] Dec 01 21:59:55 crc kubenswrapper[4857]: I1201 21:59:55.849645 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84a4e964-d1fe-48f0-a76b-ea193fac9f44" path="/var/lib/kubelet/pods/84a4e964-d1fe-48f0-a76b-ea193fac9f44/volumes" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.150362 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2"] Dec 01 22:00:00 crc kubenswrapper[4857]: E1201 22:00:00.151718 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3325d8c6-37b4-40e4-819a-8dff28a9c6c9" containerName="init" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.151742 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3325d8c6-37b4-40e4-819a-8dff28a9c6c9" containerName="init" Dec 01 22:00:00 crc kubenswrapper[4857]: E1201 22:00:00.151770 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84a4e964-d1fe-48f0-a76b-ea193fac9f44" containerName="dnsmasq-dns" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.151785 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="84a4e964-d1fe-48f0-a76b-ea193fac9f44" containerName="dnsmasq-dns" Dec 01 22:00:00 crc kubenswrapper[4857]: E1201 22:00:00.151808 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84a4e964-d1fe-48f0-a76b-ea193fac9f44" containerName="init" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.151823 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="84a4e964-d1fe-48f0-a76b-ea193fac9f44" containerName="init" Dec 01 22:00:00 crc kubenswrapper[4857]: E1201 22:00:00.151868 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3325d8c6-37b4-40e4-819a-8dff28a9c6c9" containerName="dnsmasq-dns" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.151880 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3325d8c6-37b4-40e4-819a-8dff28a9c6c9" containerName="dnsmasq-dns" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.152327 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="84a4e964-d1fe-48f0-a76b-ea193fac9f44" containerName="dnsmasq-dns" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.152361 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="3325d8c6-37b4-40e4-819a-8dff28a9c6c9" containerName="dnsmasq-dns" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.155289 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.159378 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.159797 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.187463 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2"] Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.282506 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1112bd6e-216a-48cb-9505-b1caf7d178cd-secret-volume\") pod \"collect-profiles-29410440-nl2z2\" (UID: \"1112bd6e-216a-48cb-9505-b1caf7d178cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.282656 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1112bd6e-216a-48cb-9505-b1caf7d178cd-config-volume\") pod \"collect-profiles-29410440-nl2z2\" (UID: \"1112bd6e-216a-48cb-9505-b1caf7d178cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.282723 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2bb9\" (UniqueName: \"kubernetes.io/projected/1112bd6e-216a-48cb-9505-b1caf7d178cd-kube-api-access-f2bb9\") pod \"collect-profiles-29410440-nl2z2\" (UID: \"1112bd6e-216a-48cb-9505-b1caf7d178cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.385509 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2bb9\" (UniqueName: \"kubernetes.io/projected/1112bd6e-216a-48cb-9505-b1caf7d178cd-kube-api-access-f2bb9\") pod \"collect-profiles-29410440-nl2z2\" (UID: \"1112bd6e-216a-48cb-9505-b1caf7d178cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.386215 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1112bd6e-216a-48cb-9505-b1caf7d178cd-secret-volume\") pod \"collect-profiles-29410440-nl2z2\" (UID: \"1112bd6e-216a-48cb-9505-b1caf7d178cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.386269 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1112bd6e-216a-48cb-9505-b1caf7d178cd-config-volume\") pod \"collect-profiles-29410440-nl2z2\" (UID: \"1112bd6e-216a-48cb-9505-b1caf7d178cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.387665 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1112bd6e-216a-48cb-9505-b1caf7d178cd-config-volume\") pod \"collect-profiles-29410440-nl2z2\" (UID: \"1112bd6e-216a-48cb-9505-b1caf7d178cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.397389 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1112bd6e-216a-48cb-9505-b1caf7d178cd-secret-volume\") pod \"collect-profiles-29410440-nl2z2\" (UID: \"1112bd6e-216a-48cb-9505-b1caf7d178cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.408955 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2bb9\" (UniqueName: \"kubernetes.io/projected/1112bd6e-216a-48cb-9505-b1caf7d178cd-kube-api-access-f2bb9\") pod \"collect-profiles-29410440-nl2z2\" (UID: \"1112bd6e-216a-48cb-9505-b1caf7d178cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:00 crc kubenswrapper[4857]: I1201 22:00:00.495187 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:01 crc kubenswrapper[4857]: I1201 22:00:01.029546 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2"] Dec 01 22:00:01 crc kubenswrapper[4857]: I1201 22:00:01.496060 4857 generic.go:334] "Generic (PLEG): container finished" podID="1112bd6e-216a-48cb-9505-b1caf7d178cd" containerID="d964aba06e670c29e4ba369c125f0d4a923af04e21298593213068e8a85a1bef" exitCode=0 Dec 01 22:00:01 crc kubenswrapper[4857]: I1201 22:00:01.496214 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" event={"ID":"1112bd6e-216a-48cb-9505-b1caf7d178cd","Type":"ContainerDied","Data":"d964aba06e670c29e4ba369c125f0d4a923af04e21298593213068e8a85a1bef"} Dec 01 22:00:01 crc kubenswrapper[4857]: I1201 22:00:01.496370 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" event={"ID":"1112bd6e-216a-48cb-9505-b1caf7d178cd","Type":"ContainerStarted","Data":"7178453a6fd5db35a2d7854cce3ffed491edc53391881d9185fea1477f4d67cc"} Dec 01 22:00:01 crc kubenswrapper[4857]: E1201 22:00:01.632008 4857 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4f80b30_d0e9_4c6c_ab50_7136ae3a5005.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4f80b30_d0e9_4c6c_ab50_7136ae3a5005.slice/crio-15bf320b8ef06b4c698462a3b008702e3b2b5fbea0846ccf295e751b305b75ce\": RecentStats: unable to find data in memory cache]" Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:02.999968 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:03.052596 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2bb9\" (UniqueName: \"kubernetes.io/projected/1112bd6e-216a-48cb-9505-b1caf7d178cd-kube-api-access-f2bb9\") pod \"1112bd6e-216a-48cb-9505-b1caf7d178cd\" (UID: \"1112bd6e-216a-48cb-9505-b1caf7d178cd\") " Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:03.052961 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1112bd6e-216a-48cb-9505-b1caf7d178cd-secret-volume\") pod \"1112bd6e-216a-48cb-9505-b1caf7d178cd\" (UID: \"1112bd6e-216a-48cb-9505-b1caf7d178cd\") " Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:03.053069 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1112bd6e-216a-48cb-9505-b1caf7d178cd-config-volume\") pod \"1112bd6e-216a-48cb-9505-b1caf7d178cd\" (UID: \"1112bd6e-216a-48cb-9505-b1caf7d178cd\") " Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:03.053966 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1112bd6e-216a-48cb-9505-b1caf7d178cd-config-volume" (OuterVolumeSpecName: "config-volume") pod "1112bd6e-216a-48cb-9505-b1caf7d178cd" (UID: "1112bd6e-216a-48cb-9505-b1caf7d178cd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:03.060963 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1112bd6e-216a-48cb-9505-b1caf7d178cd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1112bd6e-216a-48cb-9505-b1caf7d178cd" (UID: "1112bd6e-216a-48cb-9505-b1caf7d178cd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:03.061388 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1112bd6e-216a-48cb-9505-b1caf7d178cd-kube-api-access-f2bb9" (OuterVolumeSpecName: "kube-api-access-f2bb9") pod "1112bd6e-216a-48cb-9505-b1caf7d178cd" (UID: "1112bd6e-216a-48cb-9505-b1caf7d178cd"). InnerVolumeSpecName "kube-api-access-f2bb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:03.155450 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2bb9\" (UniqueName: \"kubernetes.io/projected/1112bd6e-216a-48cb-9505-b1caf7d178cd-kube-api-access-f2bb9\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:03.155491 4857 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1112bd6e-216a-48cb-9505-b1caf7d178cd-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:03.155503 4857 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1112bd6e-216a-48cb-9505-b1caf7d178cd-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:03.632326 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" event={"ID":"1112bd6e-216a-48cb-9505-b1caf7d178cd","Type":"ContainerDied","Data":"7178453a6fd5db35a2d7854cce3ffed491edc53391881d9185fea1477f4d67cc"} Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:03.632746 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7178453a6fd5db35a2d7854cce3ffed491edc53391881d9185fea1477f4d67cc" Dec 01 22:00:03 crc kubenswrapper[4857]: I1201 22:00:03.633076 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.136688 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h"] Dec 01 22:00:06 crc kubenswrapper[4857]: E1201 22:00:06.138352 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1112bd6e-216a-48cb-9505-b1caf7d178cd" containerName="collect-profiles" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.138388 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="1112bd6e-216a-48cb-9505-b1caf7d178cd" containerName="collect-profiles" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.138988 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="1112bd6e-216a-48cb-9505-b1caf7d178cd" containerName="collect-profiles" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.140675 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.143932 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.144019 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.144810 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.144916 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.148236 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h"] Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.221919 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc72m\" (UniqueName: \"kubernetes.io/projected/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-kube-api-access-pc72m\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.222535 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.222591 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.222760 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.324234 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.324329 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.324437 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc72m\" (UniqueName: \"kubernetes.io/projected/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-kube-api-access-pc72m\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.324497 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.331552 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.335107 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.337549 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.345633 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc72m\" (UniqueName: \"kubernetes.io/projected/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-kube-api-access-pc72m\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:06 crc kubenswrapper[4857]: I1201 22:00:06.474982 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:07 crc kubenswrapper[4857]: I1201 22:00:07.080339 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h"] Dec 01 22:00:07 crc kubenswrapper[4857]: W1201 22:00:07.136666 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc94c543_00bf_4ce3_a8cc_1ed51f8bbf3c.slice/crio-d6836254d84032120fc87fd28136408fff3978e47ff29cfdc75a237106867f62 WatchSource:0}: Error finding container d6836254d84032120fc87fd28136408fff3978e47ff29cfdc75a237106867f62: Status 404 returned error can't find the container with id d6836254d84032120fc87fd28136408fff3978e47ff29cfdc75a237106867f62 Dec 01 22:00:07 crc kubenswrapper[4857]: I1201 22:00:07.680191 4857 generic.go:334] "Generic (PLEG): container finished" podID="cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825" containerID="78c166ce31b5bcf61a3dc702f2f0983511e762ceea37f6d54b3a49288de19622" exitCode=0 Dec 01 22:00:07 crc kubenswrapper[4857]: I1201 22:00:07.680250 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825","Type":"ContainerDied","Data":"78c166ce31b5bcf61a3dc702f2f0983511e762ceea37f6d54b3a49288de19622"} Dec 01 22:00:07 crc kubenswrapper[4857]: I1201 22:00:07.684518 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" event={"ID":"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c","Type":"ContainerStarted","Data":"d6836254d84032120fc87fd28136408fff3978e47ff29cfdc75a237106867f62"} Dec 01 22:00:07 crc kubenswrapper[4857]: I1201 22:00:07.687821 4857 generic.go:334] "Generic (PLEG): container finished" podID="a49ab0a8-09f6-4bfa-9276-c89606c0030e" containerID="3aee4b029f199f947d36fe646d64d6b2344b0e5ce2abc80e1e3a8e474346b46f" exitCode=0 Dec 01 22:00:07 crc kubenswrapper[4857]: I1201 22:00:07.687863 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a49ab0a8-09f6-4bfa-9276-c89606c0030e","Type":"ContainerDied","Data":"3aee4b029f199f947d36fe646d64d6b2344b0e5ce2abc80e1e3a8e474346b46f"} Dec 01 22:00:08 crc kubenswrapper[4857]: I1201 22:00:08.708179 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a49ab0a8-09f6-4bfa-9276-c89606c0030e","Type":"ContainerStarted","Data":"1336bc88e4f9fe2fb81005240d4da6f094d00352beb1828b3b7e6f1a4875773d"} Dec 01 22:00:08 crc kubenswrapper[4857]: I1201 22:00:08.708974 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 22:00:08 crc kubenswrapper[4857]: I1201 22:00:08.711918 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825","Type":"ContainerStarted","Data":"24b5513bed7636839d62e4e60ef02c030ead1f37458082050f6394db18a3055d"} Dec 01 22:00:08 crc kubenswrapper[4857]: I1201 22:00:08.713485 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 22:00:08 crc kubenswrapper[4857]: I1201 22:00:08.748464 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.748437543 podStartE2EDuration="37.748437543s" podCreationTimestamp="2025-12-01 21:59:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 22:00:08.743692967 +0000 UTC m=+1527.233755284" watchObservedRunningTime="2025-12-01 22:00:08.748437543 +0000 UTC m=+1527.238499860" Dec 01 22:00:08 crc kubenswrapper[4857]: I1201 22:00:08.767073 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.767053828 podStartE2EDuration="37.767053828s" podCreationTimestamp="2025-12-01 21:59:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 22:00:08.763328989 +0000 UTC m=+1527.253391306" watchObservedRunningTime="2025-12-01 22:00:08.767053828 +0000 UTC m=+1527.257116145" Dec 01 22:00:10 crc kubenswrapper[4857]: I1201 22:00:10.417685 4857 scope.go:117] "RemoveContainer" containerID="08785a340aa8e4a8be90a3d1f8822a6afbed843f3a156732aa378592d9c0061f" Dec 01 22:00:11 crc kubenswrapper[4857]: E1201 22:00:11.942230 4857 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4f80b30_d0e9_4c6c_ab50_7136ae3a5005.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4f80b30_d0e9_4c6c_ab50_7136ae3a5005.slice/crio-15bf320b8ef06b4c698462a3b008702e3b2b5fbea0846ccf295e751b305b75ce\": RecentStats: unable to find data in memory cache]" Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.308241 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t28h8"] Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.310858 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.327850 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t28h8"] Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.464932 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6xz9\" (UniqueName: \"kubernetes.io/projected/912d33c9-36eb-476e-aa91-1c48315b0e80-kube-api-access-p6xz9\") pod \"community-operators-t28h8\" (UID: \"912d33c9-36eb-476e-aa91-1c48315b0e80\") " pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.465150 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/912d33c9-36eb-476e-aa91-1c48315b0e80-catalog-content\") pod \"community-operators-t28h8\" (UID: \"912d33c9-36eb-476e-aa91-1c48315b0e80\") " pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.465359 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/912d33c9-36eb-476e-aa91-1c48315b0e80-utilities\") pod \"community-operators-t28h8\" (UID: \"912d33c9-36eb-476e-aa91-1c48315b0e80\") " pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.567004 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/912d33c9-36eb-476e-aa91-1c48315b0e80-utilities\") pod \"community-operators-t28h8\" (UID: \"912d33c9-36eb-476e-aa91-1c48315b0e80\") " pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.567233 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6xz9\" (UniqueName: \"kubernetes.io/projected/912d33c9-36eb-476e-aa91-1c48315b0e80-kube-api-access-p6xz9\") pod \"community-operators-t28h8\" (UID: \"912d33c9-36eb-476e-aa91-1c48315b0e80\") " pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.567268 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/912d33c9-36eb-476e-aa91-1c48315b0e80-catalog-content\") pod \"community-operators-t28h8\" (UID: \"912d33c9-36eb-476e-aa91-1c48315b0e80\") " pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.567420 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/912d33c9-36eb-476e-aa91-1c48315b0e80-utilities\") pod \"community-operators-t28h8\" (UID: \"912d33c9-36eb-476e-aa91-1c48315b0e80\") " pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.567731 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/912d33c9-36eb-476e-aa91-1c48315b0e80-catalog-content\") pod \"community-operators-t28h8\" (UID: \"912d33c9-36eb-476e-aa91-1c48315b0e80\") " pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.626362 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6xz9\" (UniqueName: \"kubernetes.io/projected/912d33c9-36eb-476e-aa91-1c48315b0e80-kube-api-access-p6xz9\") pod \"community-operators-t28h8\" (UID: \"912d33c9-36eb-476e-aa91-1c48315b0e80\") " pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:15 crc kubenswrapper[4857]: I1201 22:00:15.649510 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:19 crc kubenswrapper[4857]: I1201 22:00:19.902892 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t28h8"] Dec 01 22:00:19 crc kubenswrapper[4857]: I1201 22:00:19.951963 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t28h8" event={"ID":"912d33c9-36eb-476e-aa91-1c48315b0e80","Type":"ContainerStarted","Data":"920c4ddd92971af8bdc62a6936d98f30074aff5ee0619ef6303088dc4a3ddbde"} Dec 01 22:00:19 crc kubenswrapper[4857]: I1201 22:00:19.954352 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" event={"ID":"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c","Type":"ContainerStarted","Data":"f498e29866520087962e78b18d359e4a2c6c53fed4da35ee5dcf5e02c5e5fdd2"} Dec 01 22:00:19 crc kubenswrapper[4857]: I1201 22:00:19.984487 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" podStartSLOduration=1.854457307 podStartE2EDuration="13.984468838s" podCreationTimestamp="2025-12-01 22:00:06 +0000 UTC" firstStartedPulling="2025-12-01 22:00:07.14082949 +0000 UTC m=+1525.630891807" lastFinishedPulling="2025-12-01 22:00:19.270841021 +0000 UTC m=+1537.760903338" observedRunningTime="2025-12-01 22:00:19.97136747 +0000 UTC m=+1538.461429787" watchObservedRunningTime="2025-12-01 22:00:19.984468838 +0000 UTC m=+1538.474531155" Dec 01 22:00:20 crc kubenswrapper[4857]: I1201 22:00:20.974706 4857 generic.go:334] "Generic (PLEG): container finished" podID="912d33c9-36eb-476e-aa91-1c48315b0e80" containerID="50743724e5daf6d55f058086b3a86432cfaed6267d1a4c5bc1def97cd7cf074c" exitCode=0 Dec 01 22:00:20 crc kubenswrapper[4857]: I1201 22:00:20.974818 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t28h8" event={"ID":"912d33c9-36eb-476e-aa91-1c48315b0e80","Type":"ContainerDied","Data":"50743724e5daf6d55f058086b3a86432cfaed6267d1a4c5bc1def97cd7cf074c"} Dec 01 22:00:21 crc kubenswrapper[4857]: I1201 22:00:21.788204 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 22:00:21 crc kubenswrapper[4857]: I1201 22:00:21.821550 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 22:00:22 crc kubenswrapper[4857]: E1201 22:00:22.215090 4857 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4f80b30_d0e9_4c6c_ab50_7136ae3a5005.slice/crio-15bf320b8ef06b4c698462a3b008702e3b2b5fbea0846ccf295e751b305b75ce\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4f80b30_d0e9_4c6c_ab50_7136ae3a5005.slice\": RecentStats: unable to find data in memory cache]" Dec 01 22:00:23 crc kubenswrapper[4857]: I1201 22:00:22.999653 4857 generic.go:334] "Generic (PLEG): container finished" podID="912d33c9-36eb-476e-aa91-1c48315b0e80" containerID="267f135e59191ca21efe26eafc5ca3ee9628e0ba9ba22407574eade5ca096ab6" exitCode=0 Dec 01 22:00:23 crc kubenswrapper[4857]: I1201 22:00:22.999713 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t28h8" event={"ID":"912d33c9-36eb-476e-aa91-1c48315b0e80","Type":"ContainerDied","Data":"267f135e59191ca21efe26eafc5ca3ee9628e0ba9ba22407574eade5ca096ab6"} Dec 01 22:00:25 crc kubenswrapper[4857]: I1201 22:00:25.105403 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t28h8" event={"ID":"912d33c9-36eb-476e-aa91-1c48315b0e80","Type":"ContainerStarted","Data":"a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e"} Dec 01 22:00:25 crc kubenswrapper[4857]: I1201 22:00:25.140948 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t28h8" podStartSLOduration=7.438350269 podStartE2EDuration="10.140917606s" podCreationTimestamp="2025-12-01 22:00:15 +0000 UTC" firstStartedPulling="2025-12-01 22:00:20.979190935 +0000 UTC m=+1539.469253292" lastFinishedPulling="2025-12-01 22:00:23.681758312 +0000 UTC m=+1542.171820629" observedRunningTime="2025-12-01 22:00:25.13394292 +0000 UTC m=+1543.624005297" watchObservedRunningTime="2025-12-01 22:00:25.140917606 +0000 UTC m=+1543.630979963" Dec 01 22:00:25 crc kubenswrapper[4857]: I1201 22:00:25.650995 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:25 crc kubenswrapper[4857]: I1201 22:00:25.651114 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:26 crc kubenswrapper[4857]: I1201 22:00:26.723260 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-t28h8" podUID="912d33c9-36eb-476e-aa91-1c48315b0e80" containerName="registry-server" probeResult="failure" output=< Dec 01 22:00:26 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Dec 01 22:00:26 crc kubenswrapper[4857]: > Dec 01 22:00:32 crc kubenswrapper[4857]: E1201 22:00:32.526997 4857 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4f80b30_d0e9_4c6c_ab50_7136ae3a5005.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4f80b30_d0e9_4c6c_ab50_7136ae3a5005.slice/crio-15bf320b8ef06b4c698462a3b008702e3b2b5fbea0846ccf295e751b305b75ce\": RecentStats: unable to find data in memory cache]" Dec 01 22:00:34 crc kubenswrapper[4857]: I1201 22:00:34.205333 4857 generic.go:334] "Generic (PLEG): container finished" podID="bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c" containerID="f498e29866520087962e78b18d359e4a2c6c53fed4da35ee5dcf5e02c5e5fdd2" exitCode=0 Dec 01 22:00:34 crc kubenswrapper[4857]: I1201 22:00:34.205393 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" event={"ID":"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c","Type":"ContainerDied","Data":"f498e29866520087962e78b18d359e4a2c6c53fed4da35ee5dcf5e02c5e5fdd2"} Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.713843 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.770023 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.798018 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.875269 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-repo-setup-combined-ca-bundle\") pod \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.875480 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-ssh-key\") pod \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.875624 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-inventory\") pod \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.875670 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc72m\" (UniqueName: \"kubernetes.io/projected/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-kube-api-access-pc72m\") pod \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\" (UID: \"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c\") " Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.885101 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c" (UID: "bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.894178 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-kube-api-access-pc72m" (OuterVolumeSpecName: "kube-api-access-pc72m") pod "bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c" (UID: "bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c"). InnerVolumeSpecName "kube-api-access-pc72m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.908382 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c" (UID: "bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.924410 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-inventory" (OuterVolumeSpecName: "inventory") pod "bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c" (UID: "bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.959931 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t28h8"] Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.984548 4857 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.984581 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.984593 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:35 crc kubenswrapper[4857]: I1201 22:00:35.984604 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc72m\" (UniqueName: \"kubernetes.io/projected/bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c-kube-api-access-pc72m\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.233147 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.233150 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h" event={"ID":"bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c","Type":"ContainerDied","Data":"d6836254d84032120fc87fd28136408fff3978e47ff29cfdc75a237106867f62"} Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.233208 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6836254d84032120fc87fd28136408fff3978e47ff29cfdc75a237106867f62" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.338235 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k"] Dec 01 22:00:36 crc kubenswrapper[4857]: E1201 22:00:36.338708 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.338735 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.339208 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.340075 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.342506 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.342702 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.343369 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.343639 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.349481 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k"] Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.392220 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmkx9\" (UniqueName: \"kubernetes.io/projected/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-kube-api-access-kmkx9\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-twm8k\" (UID: \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.392280 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-twm8k\" (UID: \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.392367 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-twm8k\" (UID: \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.494317 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmkx9\" (UniqueName: \"kubernetes.io/projected/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-kube-api-access-kmkx9\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-twm8k\" (UID: \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.494376 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-twm8k\" (UID: \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.494441 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-twm8k\" (UID: \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.499398 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-twm8k\" (UID: \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.499473 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-twm8k\" (UID: \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.515688 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmkx9\" (UniqueName: \"kubernetes.io/projected/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-kube-api-access-kmkx9\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-twm8k\" (UID: \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:36 crc kubenswrapper[4857]: I1201 22:00:36.661438 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:37 crc kubenswrapper[4857]: I1201 22:00:37.242374 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t28h8" podUID="912d33c9-36eb-476e-aa91-1c48315b0e80" containerName="registry-server" containerID="cri-o://a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e" gracePeriod=2 Dec 01 22:00:37 crc kubenswrapper[4857]: I1201 22:00:37.416585 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k"] Dec 01 22:00:37 crc kubenswrapper[4857]: W1201 22:00:37.425255 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ae0c086_76b5_44dc_ab9f_e8988b92cc02.slice/crio-b1c9a23121622c1ae0c4a7e4ab9a54b4ea256a6d8a64d969223c5efddd791286 WatchSource:0}: Error finding container b1c9a23121622c1ae0c4a7e4ab9a54b4ea256a6d8a64d969223c5efddd791286: Status 404 returned error can't find the container with id b1c9a23121622c1ae0c4a7e4ab9a54b4ea256a6d8a64d969223c5efddd791286 Dec 01 22:00:37 crc kubenswrapper[4857]: I1201 22:00:37.759729 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:37 crc kubenswrapper[4857]: I1201 22:00:37.897255 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/912d33c9-36eb-476e-aa91-1c48315b0e80-utilities\") pod \"912d33c9-36eb-476e-aa91-1c48315b0e80\" (UID: \"912d33c9-36eb-476e-aa91-1c48315b0e80\") " Dec 01 22:00:37 crc kubenswrapper[4857]: I1201 22:00:37.897560 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/912d33c9-36eb-476e-aa91-1c48315b0e80-catalog-content\") pod \"912d33c9-36eb-476e-aa91-1c48315b0e80\" (UID: \"912d33c9-36eb-476e-aa91-1c48315b0e80\") " Dec 01 22:00:37 crc kubenswrapper[4857]: I1201 22:00:37.897705 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6xz9\" (UniqueName: \"kubernetes.io/projected/912d33c9-36eb-476e-aa91-1c48315b0e80-kube-api-access-p6xz9\") pod \"912d33c9-36eb-476e-aa91-1c48315b0e80\" (UID: \"912d33c9-36eb-476e-aa91-1c48315b0e80\") " Dec 01 22:00:37 crc kubenswrapper[4857]: I1201 22:00:37.899505 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/912d33c9-36eb-476e-aa91-1c48315b0e80-utilities" (OuterVolumeSpecName: "utilities") pod "912d33c9-36eb-476e-aa91-1c48315b0e80" (UID: "912d33c9-36eb-476e-aa91-1c48315b0e80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:00:37 crc kubenswrapper[4857]: I1201 22:00:37.903026 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/912d33c9-36eb-476e-aa91-1c48315b0e80-kube-api-access-p6xz9" (OuterVolumeSpecName: "kube-api-access-p6xz9") pod "912d33c9-36eb-476e-aa91-1c48315b0e80" (UID: "912d33c9-36eb-476e-aa91-1c48315b0e80"). InnerVolumeSpecName "kube-api-access-p6xz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:00:37 crc kubenswrapper[4857]: I1201 22:00:37.941518 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/912d33c9-36eb-476e-aa91-1c48315b0e80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "912d33c9-36eb-476e-aa91-1c48315b0e80" (UID: "912d33c9-36eb-476e-aa91-1c48315b0e80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.002369 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/912d33c9-36eb-476e-aa91-1c48315b0e80-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.002435 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/912d33c9-36eb-476e-aa91-1c48315b0e80-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.002465 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6xz9\" (UniqueName: \"kubernetes.io/projected/912d33c9-36eb-476e-aa91-1c48315b0e80-kube-api-access-p6xz9\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.254758 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" event={"ID":"8ae0c086-76b5-44dc-ab9f-e8988b92cc02","Type":"ContainerStarted","Data":"ef47024994d3852b01f5f9d60a3a56e48f3f4a67785093710171d7848bcfcefc"} Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.255167 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" event={"ID":"8ae0c086-76b5-44dc-ab9f-e8988b92cc02","Type":"ContainerStarted","Data":"b1c9a23121622c1ae0c4a7e4ab9a54b4ea256a6d8a64d969223c5efddd791286"} Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.261651 4857 generic.go:334] "Generic (PLEG): container finished" podID="912d33c9-36eb-476e-aa91-1c48315b0e80" containerID="a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e" exitCode=0 Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.261744 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t28h8" event={"ID":"912d33c9-36eb-476e-aa91-1c48315b0e80","Type":"ContainerDied","Data":"a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e"} Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.261982 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t28h8" event={"ID":"912d33c9-36eb-476e-aa91-1c48315b0e80","Type":"ContainerDied","Data":"920c4ddd92971af8bdc62a6936d98f30074aff5ee0619ef6303088dc4a3ddbde"} Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.261756 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t28h8" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.262063 4857 scope.go:117] "RemoveContainer" containerID="a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.277352 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" podStartSLOduration=1.829217292 podStartE2EDuration="2.277334436s" podCreationTimestamp="2025-12-01 22:00:36 +0000 UTC" firstStartedPulling="2025-12-01 22:00:37.431645874 +0000 UTC m=+1555.921708191" lastFinishedPulling="2025-12-01 22:00:37.879763018 +0000 UTC m=+1556.369825335" observedRunningTime="2025-12-01 22:00:38.275678102 +0000 UTC m=+1556.765740439" watchObservedRunningTime="2025-12-01 22:00:38.277334436 +0000 UTC m=+1556.767396753" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.299239 4857 scope.go:117] "RemoveContainer" containerID="267f135e59191ca21efe26eafc5ca3ee9628e0ba9ba22407574eade5ca096ab6" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.319742 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t28h8"] Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.331095 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t28h8"] Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.340484 4857 scope.go:117] "RemoveContainer" containerID="50743724e5daf6d55f058086b3a86432cfaed6267d1a4c5bc1def97cd7cf074c" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.364311 4857 scope.go:117] "RemoveContainer" containerID="a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e" Dec 01 22:00:38 crc kubenswrapper[4857]: E1201 22:00:38.364846 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e\": container with ID starting with a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e not found: ID does not exist" containerID="a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.364914 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e"} err="failed to get container status \"a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e\": rpc error: code = NotFound desc = could not find container \"a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e\": container with ID starting with a3a5bb3b68c20d63a5530427dbbbbcf8600c2cf2953b1252e5377f0235a7135e not found: ID does not exist" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.364958 4857 scope.go:117] "RemoveContainer" containerID="267f135e59191ca21efe26eafc5ca3ee9628e0ba9ba22407574eade5ca096ab6" Dec 01 22:00:38 crc kubenswrapper[4857]: E1201 22:00:38.366600 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"267f135e59191ca21efe26eafc5ca3ee9628e0ba9ba22407574eade5ca096ab6\": container with ID starting with 267f135e59191ca21efe26eafc5ca3ee9628e0ba9ba22407574eade5ca096ab6 not found: ID does not exist" containerID="267f135e59191ca21efe26eafc5ca3ee9628e0ba9ba22407574eade5ca096ab6" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.366649 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"267f135e59191ca21efe26eafc5ca3ee9628e0ba9ba22407574eade5ca096ab6"} err="failed to get container status \"267f135e59191ca21efe26eafc5ca3ee9628e0ba9ba22407574eade5ca096ab6\": rpc error: code = NotFound desc = could not find container \"267f135e59191ca21efe26eafc5ca3ee9628e0ba9ba22407574eade5ca096ab6\": container with ID starting with 267f135e59191ca21efe26eafc5ca3ee9628e0ba9ba22407574eade5ca096ab6 not found: ID does not exist" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.366679 4857 scope.go:117] "RemoveContainer" containerID="50743724e5daf6d55f058086b3a86432cfaed6267d1a4c5bc1def97cd7cf074c" Dec 01 22:00:38 crc kubenswrapper[4857]: E1201 22:00:38.367395 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50743724e5daf6d55f058086b3a86432cfaed6267d1a4c5bc1def97cd7cf074c\": container with ID starting with 50743724e5daf6d55f058086b3a86432cfaed6267d1a4c5bc1def97cd7cf074c not found: ID does not exist" containerID="50743724e5daf6d55f058086b3a86432cfaed6267d1a4c5bc1def97cd7cf074c" Dec 01 22:00:38 crc kubenswrapper[4857]: I1201 22:00:38.367442 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50743724e5daf6d55f058086b3a86432cfaed6267d1a4c5bc1def97cd7cf074c"} err="failed to get container status \"50743724e5daf6d55f058086b3a86432cfaed6267d1a4c5bc1def97cd7cf074c\": rpc error: code = NotFound desc = could not find container \"50743724e5daf6d55f058086b3a86432cfaed6267d1a4c5bc1def97cd7cf074c\": container with ID starting with 50743724e5daf6d55f058086b3a86432cfaed6267d1a4c5bc1def97cd7cf074c not found: ID does not exist" Dec 01 22:00:39 crc kubenswrapper[4857]: I1201 22:00:39.850515 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="912d33c9-36eb-476e-aa91-1c48315b0e80" path="/var/lib/kubelet/pods/912d33c9-36eb-476e-aa91-1c48315b0e80/volumes" Dec 01 22:00:41 crc kubenswrapper[4857]: I1201 22:00:41.375293 4857 generic.go:334] "Generic (PLEG): container finished" podID="8ae0c086-76b5-44dc-ab9f-e8988b92cc02" containerID="ef47024994d3852b01f5f9d60a3a56e48f3f4a67785093710171d7848bcfcefc" exitCode=0 Dec 01 22:00:41 crc kubenswrapper[4857]: I1201 22:00:41.375336 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" event={"ID":"8ae0c086-76b5-44dc-ab9f-e8988b92cc02","Type":"ContainerDied","Data":"ef47024994d3852b01f5f9d60a3a56e48f3f4a67785093710171d7848bcfcefc"} Dec 01 22:00:41 crc kubenswrapper[4857]: E1201 22:00:41.869056 4857 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/3d06de2c736f3b9a7dbf6f3e55f724594e671a56a2534b6e7ad6a10f15e62430/diff" to get inode usage: stat /var/lib/containers/storage/overlay/3d06de2c736f3b9a7dbf6f3e55f724594e671a56a2534b6e7ad6a10f15e62430/diff: no such file or directory, extraDiskErr: Dec 01 22:00:42 crc kubenswrapper[4857]: I1201 22:00:42.938394 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.039483 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmkx9\" (UniqueName: \"kubernetes.io/projected/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-kube-api-access-kmkx9\") pod \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\" (UID: \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\") " Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.039636 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-inventory\") pod \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\" (UID: \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\") " Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.039675 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-ssh-key\") pod \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\" (UID: \"8ae0c086-76b5-44dc-ab9f-e8988b92cc02\") " Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.045141 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-kube-api-access-kmkx9" (OuterVolumeSpecName: "kube-api-access-kmkx9") pod "8ae0c086-76b5-44dc-ab9f-e8988b92cc02" (UID: "8ae0c086-76b5-44dc-ab9f-e8988b92cc02"). InnerVolumeSpecName "kube-api-access-kmkx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.069605 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8ae0c086-76b5-44dc-ab9f-e8988b92cc02" (UID: "8ae0c086-76b5-44dc-ab9f-e8988b92cc02"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.069635 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-inventory" (OuterVolumeSpecName: "inventory") pod "8ae0c086-76b5-44dc-ab9f-e8988b92cc02" (UID: "8ae0c086-76b5-44dc-ab9f-e8988b92cc02"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.143184 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmkx9\" (UniqueName: \"kubernetes.io/projected/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-kube-api-access-kmkx9\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.143270 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.143295 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ae0c086-76b5-44dc-ab9f-e8988b92cc02-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.417879 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" event={"ID":"8ae0c086-76b5-44dc-ab9f-e8988b92cc02","Type":"ContainerDied","Data":"b1c9a23121622c1ae0c4a7e4ab9a54b4ea256a6d8a64d969223c5efddd791286"} Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.418414 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1c9a23121622c1ae0c4a7e4ab9a54b4ea256a6d8a64d969223c5efddd791286" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.417998 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-twm8k" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.493032 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt"] Dec 01 22:00:43 crc kubenswrapper[4857]: E1201 22:00:43.493618 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae0c086-76b5-44dc-ab9f-e8988b92cc02" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.493642 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae0c086-76b5-44dc-ab9f-e8988b92cc02" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 01 22:00:43 crc kubenswrapper[4857]: E1201 22:00:43.493669 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="912d33c9-36eb-476e-aa91-1c48315b0e80" containerName="extract-utilities" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.493678 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="912d33c9-36eb-476e-aa91-1c48315b0e80" containerName="extract-utilities" Dec 01 22:00:43 crc kubenswrapper[4857]: E1201 22:00:43.493690 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="912d33c9-36eb-476e-aa91-1c48315b0e80" containerName="registry-server" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.493698 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="912d33c9-36eb-476e-aa91-1c48315b0e80" containerName="registry-server" Dec 01 22:00:43 crc kubenswrapper[4857]: E1201 22:00:43.493746 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="912d33c9-36eb-476e-aa91-1c48315b0e80" containerName="extract-content" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.493754 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="912d33c9-36eb-476e-aa91-1c48315b0e80" containerName="extract-content" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.494008 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="912d33c9-36eb-476e-aa91-1c48315b0e80" containerName="registry-server" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.494054 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae0c086-76b5-44dc-ab9f-e8988b92cc02" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.494906 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.499800 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.500026 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.500173 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.500310 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.515764 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt"] Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.653858 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.654177 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxrtp\" (UniqueName: \"kubernetes.io/projected/3916d4da-634b-4ade-be14-3a0c60e6a9d2-kube-api-access-gxrtp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.654383 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.654506 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.756108 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.756199 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.756405 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxrtp\" (UniqueName: \"kubernetes.io/projected/3916d4da-634b-4ade-be14-3a0c60e6a9d2-kube-api-access-gxrtp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.756490 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.761715 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.762478 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.767278 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.779978 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxrtp\" (UniqueName: \"kubernetes.io/projected/3916d4da-634b-4ade-be14-3a0c60e6a9d2-kube-api-access-gxrtp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:43 crc kubenswrapper[4857]: I1201 22:00:43.823012 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:00:44 crc kubenswrapper[4857]: I1201 22:00:44.257413 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt"] Dec 01 22:00:44 crc kubenswrapper[4857]: I1201 22:00:44.429290 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" event={"ID":"3916d4da-634b-4ade-be14-3a0c60e6a9d2","Type":"ContainerStarted","Data":"4d48c162f93cd9f94a2c15b2db0b83f712d7d59544c020ec0b4af8f1dc14ba28"} Dec 01 22:00:45 crc kubenswrapper[4857]: I1201 22:00:45.439465 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" event={"ID":"3916d4da-634b-4ade-be14-3a0c60e6a9d2","Type":"ContainerStarted","Data":"f5b0f3752bd0303c86fcc373ac8cadc43ef09cea589e8ba267ab89ceb3188e02"} Dec 01 22:00:45 crc kubenswrapper[4857]: I1201 22:00:45.462870 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" podStartSLOduration=1.96599586 podStartE2EDuration="2.46285099s" podCreationTimestamp="2025-12-01 22:00:43 +0000 UTC" firstStartedPulling="2025-12-01 22:00:44.257617112 +0000 UTC m=+1562.747679429" lastFinishedPulling="2025-12-01 22:00:44.754472222 +0000 UTC m=+1563.244534559" observedRunningTime="2025-12-01 22:00:45.459386947 +0000 UTC m=+1563.949449264" watchObservedRunningTime="2025-12-01 22:00:45.46285099 +0000 UTC m=+1563.952913307" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.158551 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29410441-m46zq"] Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.160661 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.195393 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410441-m46zq"] Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.276784 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-combined-ca-bundle\") pod \"keystone-cron-29410441-m46zq\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.276849 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssdxc\" (UniqueName: \"kubernetes.io/projected/a8480d29-0c86-4ee6-85ea-6be3087135ec-kube-api-access-ssdxc\") pod \"keystone-cron-29410441-m46zq\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.276955 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-fernet-keys\") pod \"keystone-cron-29410441-m46zq\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.277025 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-config-data\") pod \"keystone-cron-29410441-m46zq\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.380197 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-combined-ca-bundle\") pod \"keystone-cron-29410441-m46zq\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.380274 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssdxc\" (UniqueName: \"kubernetes.io/projected/a8480d29-0c86-4ee6-85ea-6be3087135ec-kube-api-access-ssdxc\") pod \"keystone-cron-29410441-m46zq\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.380334 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-fernet-keys\") pod \"keystone-cron-29410441-m46zq\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.380373 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-config-data\") pod \"keystone-cron-29410441-m46zq\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.391317 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-fernet-keys\") pod \"keystone-cron-29410441-m46zq\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.399382 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-config-data\") pod \"keystone-cron-29410441-m46zq\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.411003 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-combined-ca-bundle\") pod \"keystone-cron-29410441-m46zq\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.426004 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssdxc\" (UniqueName: \"kubernetes.io/projected/a8480d29-0c86-4ee6-85ea-6be3087135ec-kube-api-access-ssdxc\") pod \"keystone-cron-29410441-m46zq\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.491736 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:00 crc kubenswrapper[4857]: I1201 22:01:00.786904 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410441-m46zq"] Dec 01 22:01:01 crc kubenswrapper[4857]: I1201 22:01:01.640201 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410441-m46zq" event={"ID":"a8480d29-0c86-4ee6-85ea-6be3087135ec","Type":"ContainerStarted","Data":"959d7b4f9e3d5a21d9b85f7777e494b1e218382fe55475ed377c795ae77b5250"} Dec 01 22:01:01 crc kubenswrapper[4857]: I1201 22:01:01.640476 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410441-m46zq" event={"ID":"a8480d29-0c86-4ee6-85ea-6be3087135ec","Type":"ContainerStarted","Data":"a55650abc80999155f2e17121174e6c5ebf420b4310c997b0b0a19e09f792978"} Dec 01 22:01:03 crc kubenswrapper[4857]: I1201 22:01:03.658522 4857 generic.go:334] "Generic (PLEG): container finished" podID="a8480d29-0c86-4ee6-85ea-6be3087135ec" containerID="959d7b4f9e3d5a21d9b85f7777e494b1e218382fe55475ed377c795ae77b5250" exitCode=0 Dec 01 22:01:03 crc kubenswrapper[4857]: I1201 22:01:03.658604 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410441-m46zq" event={"ID":"a8480d29-0c86-4ee6-85ea-6be3087135ec","Type":"ContainerDied","Data":"959d7b4f9e3d5a21d9b85f7777e494b1e218382fe55475ed377c795ae77b5250"} Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.029233 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.200758 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-config-data\") pod \"a8480d29-0c86-4ee6-85ea-6be3087135ec\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.200824 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-combined-ca-bundle\") pod \"a8480d29-0c86-4ee6-85ea-6be3087135ec\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.200907 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-fernet-keys\") pod \"a8480d29-0c86-4ee6-85ea-6be3087135ec\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.201150 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssdxc\" (UniqueName: \"kubernetes.io/projected/a8480d29-0c86-4ee6-85ea-6be3087135ec-kube-api-access-ssdxc\") pod \"a8480d29-0c86-4ee6-85ea-6be3087135ec\" (UID: \"a8480d29-0c86-4ee6-85ea-6be3087135ec\") " Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.209349 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a8480d29-0c86-4ee6-85ea-6be3087135ec" (UID: "a8480d29-0c86-4ee6-85ea-6be3087135ec"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.209473 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8480d29-0c86-4ee6-85ea-6be3087135ec-kube-api-access-ssdxc" (OuterVolumeSpecName: "kube-api-access-ssdxc") pod "a8480d29-0c86-4ee6-85ea-6be3087135ec" (UID: "a8480d29-0c86-4ee6-85ea-6be3087135ec"). InnerVolumeSpecName "kube-api-access-ssdxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.235346 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8480d29-0c86-4ee6-85ea-6be3087135ec" (UID: "a8480d29-0c86-4ee6-85ea-6be3087135ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.296793 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-config-data" (OuterVolumeSpecName: "config-data") pod "a8480d29-0c86-4ee6-85ea-6be3087135ec" (UID: "a8480d29-0c86-4ee6-85ea-6be3087135ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.303753 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssdxc\" (UniqueName: \"kubernetes.io/projected/a8480d29-0c86-4ee6-85ea-6be3087135ec-kube-api-access-ssdxc\") on node \"crc\" DevicePath \"\"" Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.303934 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.304016 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.304114 4857 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8480d29-0c86-4ee6-85ea-6be3087135ec-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.679223 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410441-m46zq" event={"ID":"a8480d29-0c86-4ee6-85ea-6be3087135ec","Type":"ContainerDied","Data":"a55650abc80999155f2e17121174e6c5ebf420b4310c997b0b0a19e09f792978"} Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.679268 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a55650abc80999155f2e17121174e6c5ebf420b4310c997b0b0a19e09f792978" Dec 01 22:01:05 crc kubenswrapper[4857]: I1201 22:01:05.679285 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410441-m46zq" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.276704 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p9lz7"] Dec 01 22:01:14 crc kubenswrapper[4857]: E1201 22:01:14.277759 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8480d29-0c86-4ee6-85ea-6be3087135ec" containerName="keystone-cron" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.277776 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8480d29-0c86-4ee6-85ea-6be3087135ec" containerName="keystone-cron" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.277979 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8480d29-0c86-4ee6-85ea-6be3087135ec" containerName="keystone-cron" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.279503 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.290932 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p9lz7"] Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.409817 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2t2l\" (UniqueName: \"kubernetes.io/projected/42408880-94de-4833-9f0b-91dcd74043d6-kube-api-access-k2t2l\") pod \"certified-operators-p9lz7\" (UID: \"42408880-94de-4833-9f0b-91dcd74043d6\") " pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.409874 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42408880-94de-4833-9f0b-91dcd74043d6-utilities\") pod \"certified-operators-p9lz7\" (UID: \"42408880-94de-4833-9f0b-91dcd74043d6\") " pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.410027 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42408880-94de-4833-9f0b-91dcd74043d6-catalog-content\") pod \"certified-operators-p9lz7\" (UID: \"42408880-94de-4833-9f0b-91dcd74043d6\") " pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.512123 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2t2l\" (UniqueName: \"kubernetes.io/projected/42408880-94de-4833-9f0b-91dcd74043d6-kube-api-access-k2t2l\") pod \"certified-operators-p9lz7\" (UID: \"42408880-94de-4833-9f0b-91dcd74043d6\") " pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.512183 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42408880-94de-4833-9f0b-91dcd74043d6-utilities\") pod \"certified-operators-p9lz7\" (UID: \"42408880-94de-4833-9f0b-91dcd74043d6\") " pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.512272 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42408880-94de-4833-9f0b-91dcd74043d6-catalog-content\") pod \"certified-operators-p9lz7\" (UID: \"42408880-94de-4833-9f0b-91dcd74043d6\") " pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.512904 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42408880-94de-4833-9f0b-91dcd74043d6-utilities\") pod \"certified-operators-p9lz7\" (UID: \"42408880-94de-4833-9f0b-91dcd74043d6\") " pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.512965 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42408880-94de-4833-9f0b-91dcd74043d6-catalog-content\") pod \"certified-operators-p9lz7\" (UID: \"42408880-94de-4833-9f0b-91dcd74043d6\") " pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.542898 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2t2l\" (UniqueName: \"kubernetes.io/projected/42408880-94de-4833-9f0b-91dcd74043d6-kube-api-access-k2t2l\") pod \"certified-operators-p9lz7\" (UID: \"42408880-94de-4833-9f0b-91dcd74043d6\") " pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:14 crc kubenswrapper[4857]: I1201 22:01:14.614848 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:15 crc kubenswrapper[4857]: I1201 22:01:15.184953 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p9lz7"] Dec 01 22:01:15 crc kubenswrapper[4857]: I1201 22:01:15.799640 4857 generic.go:334] "Generic (PLEG): container finished" podID="42408880-94de-4833-9f0b-91dcd74043d6" containerID="3a36903873c4e292a03e41c4ffd8417fa239137d599beff523200204b9691d6d" exitCode=0 Dec 01 22:01:15 crc kubenswrapper[4857]: I1201 22:01:15.799712 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9lz7" event={"ID":"42408880-94de-4833-9f0b-91dcd74043d6","Type":"ContainerDied","Data":"3a36903873c4e292a03e41c4ffd8417fa239137d599beff523200204b9691d6d"} Dec 01 22:01:15 crc kubenswrapper[4857]: I1201 22:01:15.799749 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9lz7" event={"ID":"42408880-94de-4833-9f0b-91dcd74043d6","Type":"ContainerStarted","Data":"9632fe001caf748ec5cc2a0975b670b1ff5d14d75b6e464624d5ee2d9053643a"} Dec 01 22:01:18 crc kubenswrapper[4857]: I1201 22:01:18.837493 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9lz7" event={"ID":"42408880-94de-4833-9f0b-91dcd74043d6","Type":"ContainerStarted","Data":"9af6c63c6c321b3c20603e07291dcb3574ae69cc5d683d7b6bfb5c3082a70ddf"} Dec 01 22:01:19 crc kubenswrapper[4857]: I1201 22:01:19.483773 4857 scope.go:117] "RemoveContainer" containerID="d41567b7fae5b528ba83d21d75265a44c2e44823455abdf64fc298e6d51f7a33" Dec 01 22:01:19 crc kubenswrapper[4857]: I1201 22:01:19.586365 4857 scope.go:117] "RemoveContainer" containerID="7936e0e57d464af88b39cfa9e719b2e64386ed51d93bf91ea00cba4b6f4e14b9" Dec 01 22:01:19 crc kubenswrapper[4857]: I1201 22:01:19.618705 4857 scope.go:117] "RemoveContainer" containerID="c15dd48459e3aae6a3a128fbaee6c013ca37c12ae6d68aaa11e2fa2339a4f126" Dec 01 22:01:19 crc kubenswrapper[4857]: I1201 22:01:19.854970 4857 generic.go:334] "Generic (PLEG): container finished" podID="42408880-94de-4833-9f0b-91dcd74043d6" containerID="9af6c63c6c321b3c20603e07291dcb3574ae69cc5d683d7b6bfb5c3082a70ddf" exitCode=0 Dec 01 22:01:19 crc kubenswrapper[4857]: I1201 22:01:19.855968 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9lz7" event={"ID":"42408880-94de-4833-9f0b-91dcd74043d6","Type":"ContainerDied","Data":"9af6c63c6c321b3c20603e07291dcb3574ae69cc5d683d7b6bfb5c3082a70ddf"} Dec 01 22:01:20 crc kubenswrapper[4857]: I1201 22:01:20.871816 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9lz7" event={"ID":"42408880-94de-4833-9f0b-91dcd74043d6","Type":"ContainerStarted","Data":"64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde"} Dec 01 22:01:20 crc kubenswrapper[4857]: I1201 22:01:20.902133 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p9lz7" podStartSLOduration=2.329009851 podStartE2EDuration="6.902100286s" podCreationTimestamp="2025-12-01 22:01:14 +0000 UTC" firstStartedPulling="2025-12-01 22:01:15.804902897 +0000 UTC m=+1594.294965244" lastFinishedPulling="2025-12-01 22:01:20.377993322 +0000 UTC m=+1598.868055679" observedRunningTime="2025-12-01 22:01:20.897508384 +0000 UTC m=+1599.387570711" watchObservedRunningTime="2025-12-01 22:01:20.902100286 +0000 UTC m=+1599.392162643" Dec 01 22:01:24 crc kubenswrapper[4857]: I1201 22:01:24.615826 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:24 crc kubenswrapper[4857]: I1201 22:01:24.616527 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:24 crc kubenswrapper[4857]: I1201 22:01:24.689904 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:34 crc kubenswrapper[4857]: I1201 22:01:34.683285 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:34 crc kubenswrapper[4857]: I1201 22:01:34.753796 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p9lz7"] Dec 01 22:01:35 crc kubenswrapper[4857]: I1201 22:01:35.036699 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p9lz7" podUID="42408880-94de-4833-9f0b-91dcd74043d6" containerName="registry-server" containerID="cri-o://64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde" gracePeriod=2 Dec 01 22:01:35 crc kubenswrapper[4857]: I1201 22:01:35.549983 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:35 crc kubenswrapper[4857]: I1201 22:01:35.691329 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42408880-94de-4833-9f0b-91dcd74043d6-catalog-content\") pod \"42408880-94de-4833-9f0b-91dcd74043d6\" (UID: \"42408880-94de-4833-9f0b-91dcd74043d6\") " Dec 01 22:01:35 crc kubenswrapper[4857]: I1201 22:01:35.691421 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42408880-94de-4833-9f0b-91dcd74043d6-utilities\") pod \"42408880-94de-4833-9f0b-91dcd74043d6\" (UID: \"42408880-94de-4833-9f0b-91dcd74043d6\") " Dec 01 22:01:35 crc kubenswrapper[4857]: I1201 22:01:35.691458 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2t2l\" (UniqueName: \"kubernetes.io/projected/42408880-94de-4833-9f0b-91dcd74043d6-kube-api-access-k2t2l\") pod \"42408880-94de-4833-9f0b-91dcd74043d6\" (UID: \"42408880-94de-4833-9f0b-91dcd74043d6\") " Dec 01 22:01:35 crc kubenswrapper[4857]: I1201 22:01:35.693374 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42408880-94de-4833-9f0b-91dcd74043d6-utilities" (OuterVolumeSpecName: "utilities") pod "42408880-94de-4833-9f0b-91dcd74043d6" (UID: "42408880-94de-4833-9f0b-91dcd74043d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:01:35 crc kubenswrapper[4857]: I1201 22:01:35.700202 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42408880-94de-4833-9f0b-91dcd74043d6-kube-api-access-k2t2l" (OuterVolumeSpecName: "kube-api-access-k2t2l") pod "42408880-94de-4833-9f0b-91dcd74043d6" (UID: "42408880-94de-4833-9f0b-91dcd74043d6"). InnerVolumeSpecName "kube-api-access-k2t2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:01:35 crc kubenswrapper[4857]: I1201 22:01:35.749166 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42408880-94de-4833-9f0b-91dcd74043d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42408880-94de-4833-9f0b-91dcd74043d6" (UID: "42408880-94de-4833-9f0b-91dcd74043d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:01:35 crc kubenswrapper[4857]: I1201 22:01:35.793560 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42408880-94de-4833-9f0b-91dcd74043d6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:01:35 crc kubenswrapper[4857]: I1201 22:01:35.793612 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42408880-94de-4833-9f0b-91dcd74043d6-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:01:35 crc kubenswrapper[4857]: I1201 22:01:35.793622 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2t2l\" (UniqueName: \"kubernetes.io/projected/42408880-94de-4833-9f0b-91dcd74043d6-kube-api-access-k2t2l\") on node \"crc\" DevicePath \"\"" Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.049651 4857 generic.go:334] "Generic (PLEG): container finished" podID="42408880-94de-4833-9f0b-91dcd74043d6" containerID="64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde" exitCode=0 Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.049686 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9lz7" event={"ID":"42408880-94de-4833-9f0b-91dcd74043d6","Type":"ContainerDied","Data":"64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde"} Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.049723 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9lz7" Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.049750 4857 scope.go:117] "RemoveContainer" containerID="64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde" Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.049727 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9lz7" event={"ID":"42408880-94de-4833-9f0b-91dcd74043d6","Type":"ContainerDied","Data":"9632fe001caf748ec5cc2a0975b670b1ff5d14d75b6e464624d5ee2d9053643a"} Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.094267 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p9lz7"] Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.095073 4857 scope.go:117] "RemoveContainer" containerID="9af6c63c6c321b3c20603e07291dcb3574ae69cc5d683d7b6bfb5c3082a70ddf" Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.133741 4857 scope.go:117] "RemoveContainer" containerID="3a36903873c4e292a03e41c4ffd8417fa239137d599beff523200204b9691d6d" Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.138458 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p9lz7"] Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.203217 4857 scope.go:117] "RemoveContainer" containerID="64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde" Dec 01 22:01:36 crc kubenswrapper[4857]: E1201 22:01:36.203796 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde\": container with ID starting with 64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde not found: ID does not exist" containerID="64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde" Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.203832 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde"} err="failed to get container status \"64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde\": rpc error: code = NotFound desc = could not find container \"64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde\": container with ID starting with 64b1bf7fbce513c5ef30eed3ece6a971c6affe1bc19968e0e72f8bd041f4bbde not found: ID does not exist" Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.203874 4857 scope.go:117] "RemoveContainer" containerID="9af6c63c6c321b3c20603e07291dcb3574ae69cc5d683d7b6bfb5c3082a70ddf" Dec 01 22:01:36 crc kubenswrapper[4857]: E1201 22:01:36.204580 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9af6c63c6c321b3c20603e07291dcb3574ae69cc5d683d7b6bfb5c3082a70ddf\": container with ID starting with 9af6c63c6c321b3c20603e07291dcb3574ae69cc5d683d7b6bfb5c3082a70ddf not found: ID does not exist" containerID="9af6c63c6c321b3c20603e07291dcb3574ae69cc5d683d7b6bfb5c3082a70ddf" Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.204607 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9af6c63c6c321b3c20603e07291dcb3574ae69cc5d683d7b6bfb5c3082a70ddf"} err="failed to get container status \"9af6c63c6c321b3c20603e07291dcb3574ae69cc5d683d7b6bfb5c3082a70ddf\": rpc error: code = NotFound desc = could not find container \"9af6c63c6c321b3c20603e07291dcb3574ae69cc5d683d7b6bfb5c3082a70ddf\": container with ID starting with 9af6c63c6c321b3c20603e07291dcb3574ae69cc5d683d7b6bfb5c3082a70ddf not found: ID does not exist" Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.204624 4857 scope.go:117] "RemoveContainer" containerID="3a36903873c4e292a03e41c4ffd8417fa239137d599beff523200204b9691d6d" Dec 01 22:01:36 crc kubenswrapper[4857]: E1201 22:01:36.204973 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a36903873c4e292a03e41c4ffd8417fa239137d599beff523200204b9691d6d\": container with ID starting with 3a36903873c4e292a03e41c4ffd8417fa239137d599beff523200204b9691d6d not found: ID does not exist" containerID="3a36903873c4e292a03e41c4ffd8417fa239137d599beff523200204b9691d6d" Dec 01 22:01:36 crc kubenswrapper[4857]: I1201 22:01:36.205002 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a36903873c4e292a03e41c4ffd8417fa239137d599beff523200204b9691d6d"} err="failed to get container status \"3a36903873c4e292a03e41c4ffd8417fa239137d599beff523200204b9691d6d\": rpc error: code = NotFound desc = could not find container \"3a36903873c4e292a03e41c4ffd8417fa239137d599beff523200204b9691d6d\": container with ID starting with 3a36903873c4e292a03e41c4ffd8417fa239137d599beff523200204b9691d6d not found: ID does not exist" Dec 01 22:01:37 crc kubenswrapper[4857]: I1201 22:01:37.855507 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42408880-94de-4833-9f0b-91dcd74043d6" path="/var/lib/kubelet/pods/42408880-94de-4833-9f0b-91dcd74043d6/volumes" Dec 01 22:01:57 crc kubenswrapper[4857]: I1201 22:01:57.806580 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:01:57 crc kubenswrapper[4857]: I1201 22:01:57.807344 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:02:19 crc kubenswrapper[4857]: I1201 22:02:19.728169 4857 scope.go:117] "RemoveContainer" containerID="0761be5c409a673752c97c6527eca4e0dabc3ce5b2e9374fa9db1ff742f67ccd" Dec 01 22:02:19 crc kubenswrapper[4857]: I1201 22:02:19.776902 4857 scope.go:117] "RemoveContainer" containerID="152968eff1600d4b0bb4c27e45c9f98e2a9426bf72c50fed47e7d7d5af868bf2" Dec 01 22:02:19 crc kubenswrapper[4857]: I1201 22:02:19.813254 4857 scope.go:117] "RemoveContainer" containerID="bd399d9bc3979936cbf901415e37c64434e266d2e56e67201d609de9a207d0b6" Dec 01 22:02:19 crc kubenswrapper[4857]: I1201 22:02:19.844789 4857 scope.go:117] "RemoveContainer" containerID="1316489558d6e4f55dd048e49f02f63980aa08f4393838910026d208c03438eb" Dec 01 22:02:27 crc kubenswrapper[4857]: I1201 22:02:27.806932 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:02:27 crc kubenswrapper[4857]: I1201 22:02:27.807869 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:02:57 crc kubenswrapper[4857]: I1201 22:02:57.806702 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:02:57 crc kubenswrapper[4857]: I1201 22:02:57.807495 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:02:57 crc kubenswrapper[4857]: I1201 22:02:57.807567 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 22:02:57 crc kubenswrapper[4857]: I1201 22:02:57.808661 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 22:02:57 crc kubenswrapper[4857]: I1201 22:02:57.808754 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" gracePeriod=600 Dec 01 22:02:57 crc kubenswrapper[4857]: E1201 22:02:57.936323 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:02:58 crc kubenswrapper[4857]: I1201 22:02:58.267766 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" exitCode=0 Dec 01 22:02:58 crc kubenswrapper[4857]: I1201 22:02:58.268007 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35"} Dec 01 22:02:58 crc kubenswrapper[4857]: I1201 22:02:58.268448 4857 scope.go:117] "RemoveContainer" containerID="022dacb43671109561ed218fe31f14692c5b447676b12b6fcdbccdc338c9650c" Dec 01 22:02:58 crc kubenswrapper[4857]: I1201 22:02:58.269386 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:02:58 crc kubenswrapper[4857]: E1201 22:02:58.269840 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:03:09 crc kubenswrapper[4857]: I1201 22:03:09.836700 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:03:09 crc kubenswrapper[4857]: E1201 22:03:09.838967 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:03:21 crc kubenswrapper[4857]: I1201 22:03:21.850005 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:03:21 crc kubenswrapper[4857]: E1201 22:03:21.851137 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:03:35 crc kubenswrapper[4857]: I1201 22:03:35.835548 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:03:35 crc kubenswrapper[4857]: E1201 22:03:35.836351 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:03:45 crc kubenswrapper[4857]: I1201 22:03:45.070328 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-sjprl"] Dec 01 22:03:45 crc kubenswrapper[4857]: I1201 22:03:45.084984 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-sjprl"] Dec 01 22:03:45 crc kubenswrapper[4857]: I1201 22:03:45.913714 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1dcab1d-428a-402a-a901-4d0d07464487" path="/var/lib/kubelet/pods/e1dcab1d-428a-402a-a901-4d0d07464487/volumes" Dec 01 22:03:46 crc kubenswrapper[4857]: I1201 22:03:46.053428 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8640-account-create-update-chdlg"] Dec 01 22:03:46 crc kubenswrapper[4857]: I1201 22:03:46.068350 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8640-account-create-update-chdlg"] Dec 01 22:03:46 crc kubenswrapper[4857]: I1201 22:03:46.835236 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:03:46 crc kubenswrapper[4857]: E1201 22:03:46.835761 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:03:47 crc kubenswrapper[4857]: I1201 22:03:47.052014 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-qlvz9"] Dec 01 22:03:47 crc kubenswrapper[4857]: I1201 22:03:47.066807 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-qlvz9"] Dec 01 22:03:47 crc kubenswrapper[4857]: I1201 22:03:47.854746 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f" path="/var/lib/kubelet/pods/5d4e2fc3-cc9e-475b-a1fa-bb12a40e834f/volumes" Dec 01 22:03:47 crc kubenswrapper[4857]: I1201 22:03:47.856673 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b06f96e5-3ee7-4b8f-95d0-968855414abe" path="/var/lib/kubelet/pods/b06f96e5-3ee7-4b8f-95d0-968855414abe/volumes" Dec 01 22:03:48 crc kubenswrapper[4857]: I1201 22:03:48.060264 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-75e9-account-create-update-lqsf5"] Dec 01 22:03:48 crc kubenswrapper[4857]: I1201 22:03:48.077691 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-75e9-account-create-update-lqsf5"] Dec 01 22:03:49 crc kubenswrapper[4857]: I1201 22:03:49.855035 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31eea916-82d4-4918-8577-68276ba53766" path="/var/lib/kubelet/pods/31eea916-82d4-4918-8577-68276ba53766/volumes" Dec 01 22:03:51 crc kubenswrapper[4857]: I1201 22:03:51.039491 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-5zzh6"] Dec 01 22:03:51 crc kubenswrapper[4857]: I1201 22:03:51.053881 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-5zzh6"] Dec 01 22:03:51 crc kubenswrapper[4857]: I1201 22:03:51.853294 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55e4213d-69d8-4f02-adcb-e2249fb3dccf" path="/var/lib/kubelet/pods/55e4213d-69d8-4f02-adcb-e2249fb3dccf/volumes" Dec 01 22:03:52 crc kubenswrapper[4857]: I1201 22:03:52.047296 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-41ef-account-create-update-rfn85"] Dec 01 22:03:52 crc kubenswrapper[4857]: I1201 22:03:52.066596 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-41ef-account-create-update-rfn85"] Dec 01 22:03:53 crc kubenswrapper[4857]: I1201 22:03:53.855375 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6fbba64-3540-42d5-a1dd-c5e65d4e1330" path="/var/lib/kubelet/pods/b6fbba64-3540-42d5-a1dd-c5e65d4e1330/volumes" Dec 01 22:03:57 crc kubenswrapper[4857]: I1201 22:03:57.836515 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:03:57 crc kubenswrapper[4857]: E1201 22:03:57.837926 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:04:08 crc kubenswrapper[4857]: I1201 22:04:08.834993 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:04:08 crc kubenswrapper[4857]: E1201 22:04:08.836697 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:04:13 crc kubenswrapper[4857]: I1201 22:04:13.238206 4857 generic.go:334] "Generic (PLEG): container finished" podID="3916d4da-634b-4ade-be14-3a0c60e6a9d2" containerID="f5b0f3752bd0303c86fcc373ac8cadc43ef09cea589e8ba267ab89ceb3188e02" exitCode=0 Dec 01 22:04:13 crc kubenswrapper[4857]: I1201 22:04:13.238294 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" event={"ID":"3916d4da-634b-4ade-be14-3a0c60e6a9d2","Type":"ContainerDied","Data":"f5b0f3752bd0303c86fcc373ac8cadc43ef09cea589e8ba267ab89ceb3188e02"} Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.733415 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.854218 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxrtp\" (UniqueName: \"kubernetes.io/projected/3916d4da-634b-4ade-be14-3a0c60e6a9d2-kube-api-access-gxrtp\") pod \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.854378 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-ssh-key\") pod \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.854718 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-bootstrap-combined-ca-bundle\") pod \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.855359 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-inventory\") pod \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\" (UID: \"3916d4da-634b-4ade-be14-3a0c60e6a9d2\") " Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.861145 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3916d4da-634b-4ade-be14-3a0c60e6a9d2-kube-api-access-gxrtp" (OuterVolumeSpecName: "kube-api-access-gxrtp") pod "3916d4da-634b-4ade-be14-3a0c60e6a9d2" (UID: "3916d4da-634b-4ade-be14-3a0c60e6a9d2"). InnerVolumeSpecName "kube-api-access-gxrtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.861879 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3916d4da-634b-4ade-be14-3a0c60e6a9d2" (UID: "3916d4da-634b-4ade-be14-3a0c60e6a9d2"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.889245 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3916d4da-634b-4ade-be14-3a0c60e6a9d2" (UID: "3916d4da-634b-4ade-be14-3a0c60e6a9d2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.890845 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-inventory" (OuterVolumeSpecName: "inventory") pod "3916d4da-634b-4ade-be14-3a0c60e6a9d2" (UID: "3916d4da-634b-4ade-be14-3a0c60e6a9d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.957688 4857 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.957720 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.957730 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxrtp\" (UniqueName: \"kubernetes.io/projected/3916d4da-634b-4ade-be14-3a0c60e6a9d2-kube-api-access-gxrtp\") on node \"crc\" DevicePath \"\"" Dec 01 22:04:14 crc kubenswrapper[4857]: I1201 22:04:14.957740 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3916d4da-634b-4ade-be14-3a0c60e6a9d2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.270245 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" event={"ID":"3916d4da-634b-4ade-be14-3a0c60e6a9d2","Type":"ContainerDied","Data":"4d48c162f93cd9f94a2c15b2db0b83f712d7d59544c020ec0b4af8f1dc14ba28"} Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.270645 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d48c162f93cd9f94a2c15b2db0b83f712d7d59544c020ec0b4af8f1dc14ba28" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.270720 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.363302 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2"] Dec 01 22:04:15 crc kubenswrapper[4857]: E1201 22:04:15.363798 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42408880-94de-4833-9f0b-91dcd74043d6" containerName="extract-utilities" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.363816 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="42408880-94de-4833-9f0b-91dcd74043d6" containerName="extract-utilities" Dec 01 22:04:15 crc kubenswrapper[4857]: E1201 22:04:15.363825 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3916d4da-634b-4ade-be14-3a0c60e6a9d2" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.363833 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3916d4da-634b-4ade-be14-3a0c60e6a9d2" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 22:04:15 crc kubenswrapper[4857]: E1201 22:04:15.363856 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42408880-94de-4833-9f0b-91dcd74043d6" containerName="extract-content" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.363862 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="42408880-94de-4833-9f0b-91dcd74043d6" containerName="extract-content" Dec 01 22:04:15 crc kubenswrapper[4857]: E1201 22:04:15.363880 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42408880-94de-4833-9f0b-91dcd74043d6" containerName="registry-server" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.363886 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="42408880-94de-4833-9f0b-91dcd74043d6" containerName="registry-server" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.364105 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="42408880-94de-4833-9f0b-91dcd74043d6" containerName="registry-server" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.364119 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="3916d4da-634b-4ade-be14-3a0c60e6a9d2" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.364812 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.374584 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2"] Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.404130 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.404207 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.404338 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.404151 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.466672 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d419ce3b-db4a-4224-a322-69aaa24650fa-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2\" (UID: \"d419ce3b-db4a-4224-a322-69aaa24650fa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.466724 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d419ce3b-db4a-4224-a322-69aaa24650fa-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2\" (UID: \"d419ce3b-db4a-4224-a322-69aaa24650fa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.466748 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6vr4\" (UniqueName: \"kubernetes.io/projected/d419ce3b-db4a-4224-a322-69aaa24650fa-kube-api-access-x6vr4\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2\" (UID: \"d419ce3b-db4a-4224-a322-69aaa24650fa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.569020 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d419ce3b-db4a-4224-a322-69aaa24650fa-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2\" (UID: \"d419ce3b-db4a-4224-a322-69aaa24650fa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.569155 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d419ce3b-db4a-4224-a322-69aaa24650fa-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2\" (UID: \"d419ce3b-db4a-4224-a322-69aaa24650fa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.569203 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6vr4\" (UniqueName: \"kubernetes.io/projected/d419ce3b-db4a-4224-a322-69aaa24650fa-kube-api-access-x6vr4\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2\" (UID: \"d419ce3b-db4a-4224-a322-69aaa24650fa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.573164 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d419ce3b-db4a-4224-a322-69aaa24650fa-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2\" (UID: \"d419ce3b-db4a-4224-a322-69aaa24650fa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.576869 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d419ce3b-db4a-4224-a322-69aaa24650fa-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2\" (UID: \"d419ce3b-db4a-4224-a322-69aaa24650fa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.589394 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6vr4\" (UniqueName: \"kubernetes.io/projected/d419ce3b-db4a-4224-a322-69aaa24650fa-kube-api-access-x6vr4\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2\" (UID: \"d419ce3b-db4a-4224-a322-69aaa24650fa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:04:15 crc kubenswrapper[4857]: I1201 22:04:15.725858 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:04:16 crc kubenswrapper[4857]: I1201 22:04:16.410466 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2"] Dec 01 22:04:16 crc kubenswrapper[4857]: W1201 22:04:16.418400 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd419ce3b_db4a_4224_a322_69aaa24650fa.slice/crio-5654a77eca98c29cbff7fbd88c290bf4f9f4dc39ac35be783fd6855281f36d7a WatchSource:0}: Error finding container 5654a77eca98c29cbff7fbd88c290bf4f9f4dc39ac35be783fd6855281f36d7a: Status 404 returned error can't find the container with id 5654a77eca98c29cbff7fbd88c290bf4f9f4dc39ac35be783fd6855281f36d7a Dec 01 22:04:16 crc kubenswrapper[4857]: I1201 22:04:16.423295 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 22:04:17 crc kubenswrapper[4857]: I1201 22:04:17.314988 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" event={"ID":"d419ce3b-db4a-4224-a322-69aaa24650fa","Type":"ContainerStarted","Data":"5654a77eca98c29cbff7fbd88c290bf4f9f4dc39ac35be783fd6855281f36d7a"} Dec 01 22:04:18 crc kubenswrapper[4857]: I1201 22:04:18.325937 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" event={"ID":"d419ce3b-db4a-4224-a322-69aaa24650fa","Type":"ContainerStarted","Data":"39227f0c13589be3a079031db1a638bf9655a2918a2e9700e0b5a05a6ac826c2"} Dec 01 22:04:18 crc kubenswrapper[4857]: I1201 22:04:18.351910 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" podStartSLOduration=2.488432339 podStartE2EDuration="3.351889769s" podCreationTimestamp="2025-12-01 22:04:15 +0000 UTC" firstStartedPulling="2025-12-01 22:04:16.422994973 +0000 UTC m=+1774.913057300" lastFinishedPulling="2025-12-01 22:04:17.286452373 +0000 UTC m=+1775.776514730" observedRunningTime="2025-12-01 22:04:18.348432354 +0000 UTC m=+1776.838494711" watchObservedRunningTime="2025-12-01 22:04:18.351889769 +0000 UTC m=+1776.841952096" Dec 01 22:04:19 crc kubenswrapper[4857]: I1201 22:04:19.835027 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:04:19 crc kubenswrapper[4857]: E1201 22:04:19.835547 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:04:19 crc kubenswrapper[4857]: I1201 22:04:19.983244 4857 scope.go:117] "RemoveContainer" containerID="ca81184d8dca1b486457e141f51b17710fc9c8e1b9b3d57ea5b58971f2fc3a6b" Dec 01 22:04:20 crc kubenswrapper[4857]: I1201 22:04:20.023686 4857 scope.go:117] "RemoveContainer" containerID="2361a36edb8bf39369dbc352f577e586f8366a33a92ad0d587cced04acc51a77" Dec 01 22:04:20 crc kubenswrapper[4857]: I1201 22:04:20.083887 4857 scope.go:117] "RemoveContainer" containerID="2b6870dfc6a6aa4776bad120ab2416e26e4267b1479e1694092999e8e19d0a5d" Dec 01 22:04:20 crc kubenswrapper[4857]: I1201 22:04:20.140705 4857 scope.go:117] "RemoveContainer" containerID="0c0f19e0917b3adb587ed1cee632a48dc1f82d6382e639315247141964c03f6b" Dec 01 22:04:20 crc kubenswrapper[4857]: I1201 22:04:20.194753 4857 scope.go:117] "RemoveContainer" containerID="3931aa3a7a90531868722bb8b84722451b0a11cc16dd4fdc24674887903a3cd3" Dec 01 22:04:20 crc kubenswrapper[4857]: I1201 22:04:20.228022 4857 scope.go:117] "RemoveContainer" containerID="4306ceb75c17d7dafa76df7b3e43f7eaf6d2b959b17e9aaca62e1e773f1b76a6" Dec 01 22:04:20 crc kubenswrapper[4857]: I1201 22:04:20.260810 4857 scope.go:117] "RemoveContainer" containerID="9a93c47608aaad7c765c2674da01cc28398fbd69452fd420bf3fd191c83b13fc" Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.060141 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-zdjss"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.069860 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-fac7-account-create-update-h6hzj"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.077690 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-qm6hg"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.086858 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-27ab-account-create-update-4mr2x"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.100314 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-9k4pr"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.108488 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-a859-account-create-update-z4slz"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.119193 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-2fb1-account-create-update-66k42"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.128916 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-v2h7r"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.145439 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-fac7-account-create-update-h6hzj"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.156095 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-zdjss"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.163848 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-27ab-account-create-update-4mr2x"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.171775 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-v2h7r"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.179126 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-qm6hg"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.185543 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-2fb1-account-create-update-66k42"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.193175 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-9k4pr"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.199693 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-a859-account-create-update-z4slz"] Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.858840 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="018d4544-4c86-45b0-a320-77d857f6e49b" path="/var/lib/kubelet/pods/018d4544-4c86-45b0-a320-77d857f6e49b/volumes" Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.860664 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31e91520-3698-4c0a-82f7-9792c769f9f3" path="/var/lib/kubelet/pods/31e91520-3698-4c0a-82f7-9792c769f9f3/volumes" Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.862570 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c8567c2-527e-4cba-9d7b-e3da31d2e0ec" path="/var/lib/kubelet/pods/5c8567c2-527e-4cba-9d7b-e3da31d2e0ec/volumes" Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.864181 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8107394e-a9a2-43a6-bf43-3c785c0455c6" path="/var/lib/kubelet/pods/8107394e-a9a2-43a6-bf43-3c785c0455c6/volumes" Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.867159 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85b66a20-d9b6-45ca-889b-87121ebac3f0" path="/var/lib/kubelet/pods/85b66a20-d9b6-45ca-889b-87121ebac3f0/volumes" Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.868834 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87ab7c0d-04a8-4527-966d-d48abfc3c645" path="/var/lib/kubelet/pods/87ab7c0d-04a8-4527-966d-d48abfc3c645/volumes" Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.870998 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c0aebf-0263-4947-88ee-70a119ef3453" path="/var/lib/kubelet/pods/a1c0aebf-0263-4947-88ee-70a119ef3453/volumes" Dec 01 22:04:23 crc kubenswrapper[4857]: I1201 22:04:23.875803 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6d47cc8-65ed-4b57-abb2-2216c9ed451a" path="/var/lib/kubelet/pods/a6d47cc8-65ed-4b57-abb2-2216c9ed451a/volumes" Dec 01 22:04:30 crc kubenswrapper[4857]: I1201 22:04:30.073817 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-9gzqb"] Dec 01 22:04:30 crc kubenswrapper[4857]: I1201 22:04:30.093740 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-9gzqb"] Dec 01 22:04:30 crc kubenswrapper[4857]: I1201 22:04:30.877017 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:04:30 crc kubenswrapper[4857]: E1201 22:04:30.877338 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:04:31 crc kubenswrapper[4857]: I1201 22:04:31.847658 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dab7e48a-a3a1-4e9a-bfee-0e0fa712adab" path="/var/lib/kubelet/pods/dab7e48a-a3a1-4e9a-bfee-0e0fa712adab/volumes" Dec 01 22:04:41 crc kubenswrapper[4857]: I1201 22:04:41.847846 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:04:41 crc kubenswrapper[4857]: E1201 22:04:41.849530 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:04:56 crc kubenswrapper[4857]: I1201 22:04:56.835352 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:04:56 crc kubenswrapper[4857]: E1201 22:04:56.836365 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:05:10 crc kubenswrapper[4857]: I1201 22:05:10.835612 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:05:10 crc kubenswrapper[4857]: E1201 22:05:10.836667 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:05:16 crc kubenswrapper[4857]: I1201 22:05:16.046507 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nq6w4"] Dec 01 22:05:16 crc kubenswrapper[4857]: I1201 22:05:16.057622 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-9f69d"] Dec 01 22:05:16 crc kubenswrapper[4857]: I1201 22:05:16.068323 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nq6w4"] Dec 01 22:05:16 crc kubenswrapper[4857]: I1201 22:05:16.076913 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-9f69d"] Dec 01 22:05:17 crc kubenswrapper[4857]: I1201 22:05:17.050549 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-pkbkh"] Dec 01 22:05:17 crc kubenswrapper[4857]: I1201 22:05:17.064669 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-pkbkh"] Dec 01 22:05:17 crc kubenswrapper[4857]: I1201 22:05:17.850128 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30bec9e1-3c58-400b-86fe-0ddcdde2bf8b" path="/var/lib/kubelet/pods/30bec9e1-3c58-400b-86fe-0ddcdde2bf8b/volumes" Dec 01 22:05:17 crc kubenswrapper[4857]: I1201 22:05:17.851539 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54a6f0ca-e3e9-47df-bca6-3e4287d5c32e" path="/var/lib/kubelet/pods/54a6f0ca-e3e9-47df-bca6-3e4287d5c32e/volumes" Dec 01 22:05:17 crc kubenswrapper[4857]: I1201 22:05:17.852535 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f985748e-24cd-4d1b-abe8-7a579aee87af" path="/var/lib/kubelet/pods/f985748e-24cd-4d1b-abe8-7a579aee87af/volumes" Dec 01 22:05:20 crc kubenswrapper[4857]: I1201 22:05:20.433873 4857 scope.go:117] "RemoveContainer" containerID="3ade7b8b1e44a17f41d437ae925738bf97369c91a5e3f1c839100e210b78bbf7" Dec 01 22:05:20 crc kubenswrapper[4857]: I1201 22:05:20.508903 4857 scope.go:117] "RemoveContainer" containerID="df651ca2cdaef9cd85a3e4cbe423f1711e4070912bb6df2273c3182a1c58435a" Dec 01 22:05:20 crc kubenswrapper[4857]: I1201 22:05:20.540431 4857 scope.go:117] "RemoveContainer" containerID="bde0f719f33acb5825d68ec72a778784837af0491876344f6ef13c21fdcce410" Dec 01 22:05:20 crc kubenswrapper[4857]: I1201 22:05:20.588097 4857 scope.go:117] "RemoveContainer" containerID="163bcfe0cfa5f6ce6ad5cec5512dc20820896e37abca9bba3bb2273e44a3730f" Dec 01 22:05:20 crc kubenswrapper[4857]: I1201 22:05:20.634175 4857 scope.go:117] "RemoveContainer" containerID="3d7cc0a93806a18ef5ae6046437191792bd0c4c3a7bbf3d271e09fe3a296ab21" Dec 01 22:05:20 crc kubenswrapper[4857]: I1201 22:05:20.663244 4857 scope.go:117] "RemoveContainer" containerID="a48b2165996854d387a711a39f4527ae5f5798c6bc0cf92c2d60d20cb4d3a940" Dec 01 22:05:20 crc kubenswrapper[4857]: I1201 22:05:20.697450 4857 scope.go:117] "RemoveContainer" containerID="dd0faf071d58c6a335f99b92b025df88c87a5dba8a2a7200805f355393b1b86f" Dec 01 22:05:20 crc kubenswrapper[4857]: I1201 22:05:20.743449 4857 scope.go:117] "RemoveContainer" containerID="bcd3c4cc1fa08e49a17ac0058293eb0a8d38f872be03850ee36db36448ffc2e5" Dec 01 22:05:20 crc kubenswrapper[4857]: I1201 22:05:20.772428 4857 scope.go:117] "RemoveContainer" containerID="b0697fa87148ec15461681265d17e3c82144f730492dfcb10678b693aac6413e" Dec 01 22:05:20 crc kubenswrapper[4857]: I1201 22:05:20.794452 4857 scope.go:117] "RemoveContainer" containerID="eb1ba56055913886f8724efae2010400bc17d5261fd0374e49c8ae921124535c" Dec 01 22:05:20 crc kubenswrapper[4857]: I1201 22:05:20.833601 4857 scope.go:117] "RemoveContainer" containerID="84f9de00d10ea8061abd6045a4921193ab96a535804ec1b7ad5ad735a288be2b" Dec 01 22:05:20 crc kubenswrapper[4857]: I1201 22:05:20.860213 4857 scope.go:117] "RemoveContainer" containerID="7450f895776482d0d31ffe959649c74c1d508b75293a44a04f13a71ef0ca3ed5" Dec 01 22:05:21 crc kubenswrapper[4857]: I1201 22:05:21.042098 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-s58fk"] Dec 01 22:05:21 crc kubenswrapper[4857]: I1201 22:05:21.057421 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-s58fk"] Dec 01 22:05:21 crc kubenswrapper[4857]: I1201 22:05:21.846675 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:05:21 crc kubenswrapper[4857]: E1201 22:05:21.849286 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:05:21 crc kubenswrapper[4857]: I1201 22:05:21.868925 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dc6c548-e991-444e-a627-67896bf3e307" path="/var/lib/kubelet/pods/8dc6c548-e991-444e-a627-67896bf3e307/volumes" Dec 01 22:05:31 crc kubenswrapper[4857]: I1201 22:05:31.084852 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-xzd26"] Dec 01 22:05:31 crc kubenswrapper[4857]: I1201 22:05:31.102451 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-xzd26"] Dec 01 22:05:31 crc kubenswrapper[4857]: I1201 22:05:31.844728 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa7ace20-c228-447f-98b7-dc09e562d5b4" path="/var/lib/kubelet/pods/aa7ace20-c228-447f-98b7-dc09e562d5b4/volumes" Dec 01 22:05:32 crc kubenswrapper[4857]: I1201 22:05:32.038169 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-ppkc6"] Dec 01 22:05:32 crc kubenswrapper[4857]: I1201 22:05:32.147569 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-ppkc6"] Dec 01 22:05:33 crc kubenswrapper[4857]: I1201 22:05:33.846574 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66c9aa39-8cad-4807-98c7-ee329d87bb20" path="/var/lib/kubelet/pods/66c9aa39-8cad-4807-98c7-ee329d87bb20/volumes" Dec 01 22:05:35 crc kubenswrapper[4857]: I1201 22:05:35.836097 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:05:35 crc kubenswrapper[4857]: E1201 22:05:35.836513 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:05:42 crc kubenswrapper[4857]: I1201 22:05:42.039137 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-x296d"] Dec 01 22:05:42 crc kubenswrapper[4857]: I1201 22:05:42.052374 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-x296d"] Dec 01 22:05:43 crc kubenswrapper[4857]: I1201 22:05:43.846412 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc87da26-cfe1-401c-b115-bab361f3c126" path="/var/lib/kubelet/pods/dc87da26-cfe1-401c-b115-bab361f3c126/volumes" Dec 01 22:05:46 crc kubenswrapper[4857]: I1201 22:05:46.835552 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:05:46 crc kubenswrapper[4857]: E1201 22:05:46.836442 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:05:55 crc kubenswrapper[4857]: I1201 22:05:55.485790 4857 generic.go:334] "Generic (PLEG): container finished" podID="d419ce3b-db4a-4224-a322-69aaa24650fa" containerID="39227f0c13589be3a079031db1a638bf9655a2918a2e9700e0b5a05a6ac826c2" exitCode=0 Dec 01 22:05:55 crc kubenswrapper[4857]: I1201 22:05:55.485911 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" event={"ID":"d419ce3b-db4a-4224-a322-69aaa24650fa","Type":"ContainerDied","Data":"39227f0c13589be3a079031db1a638bf9655a2918a2e9700e0b5a05a6ac826c2"} Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.085166 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.142023 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6vr4\" (UniqueName: \"kubernetes.io/projected/d419ce3b-db4a-4224-a322-69aaa24650fa-kube-api-access-x6vr4\") pod \"d419ce3b-db4a-4224-a322-69aaa24650fa\" (UID: \"d419ce3b-db4a-4224-a322-69aaa24650fa\") " Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.142306 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d419ce3b-db4a-4224-a322-69aaa24650fa-ssh-key\") pod \"d419ce3b-db4a-4224-a322-69aaa24650fa\" (UID: \"d419ce3b-db4a-4224-a322-69aaa24650fa\") " Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.142402 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d419ce3b-db4a-4224-a322-69aaa24650fa-inventory\") pod \"d419ce3b-db4a-4224-a322-69aaa24650fa\" (UID: \"d419ce3b-db4a-4224-a322-69aaa24650fa\") " Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.157311 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d419ce3b-db4a-4224-a322-69aaa24650fa-kube-api-access-x6vr4" (OuterVolumeSpecName: "kube-api-access-x6vr4") pod "d419ce3b-db4a-4224-a322-69aaa24650fa" (UID: "d419ce3b-db4a-4224-a322-69aaa24650fa"). InnerVolumeSpecName "kube-api-access-x6vr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.183236 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d419ce3b-db4a-4224-a322-69aaa24650fa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d419ce3b-db4a-4224-a322-69aaa24650fa" (UID: "d419ce3b-db4a-4224-a322-69aaa24650fa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.202487 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d419ce3b-db4a-4224-a322-69aaa24650fa-inventory" (OuterVolumeSpecName: "inventory") pod "d419ce3b-db4a-4224-a322-69aaa24650fa" (UID: "d419ce3b-db4a-4224-a322-69aaa24650fa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.244215 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d419ce3b-db4a-4224-a322-69aaa24650fa-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.244257 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6vr4\" (UniqueName: \"kubernetes.io/projected/d419ce3b-db4a-4224-a322-69aaa24650fa-kube-api-access-x6vr4\") on node \"crc\" DevicePath \"\"" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.244275 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d419ce3b-db4a-4224-a322-69aaa24650fa-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.511207 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" event={"ID":"d419ce3b-db4a-4224-a322-69aaa24650fa","Type":"ContainerDied","Data":"5654a77eca98c29cbff7fbd88c290bf4f9f4dc39ac35be783fd6855281f36d7a"} Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.511272 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5654a77eca98c29cbff7fbd88c290bf4f9f4dc39ac35be783fd6855281f36d7a" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.511335 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.638957 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm"] Dec 01 22:05:57 crc kubenswrapper[4857]: E1201 22:05:57.639385 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d419ce3b-db4a-4224-a322-69aaa24650fa" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.639408 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d419ce3b-db4a-4224-a322-69aaa24650fa" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.641230 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="d419ce3b-db4a-4224-a322-69aaa24650fa" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.642168 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.645535 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.645847 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.645897 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.649286 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm"] Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.649802 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.651030 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjmbb\" (UniqueName: \"kubernetes.io/projected/0a768b62-44a4-4493-bd0d-2a536c55370d-kube-api-access-kjmbb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm\" (UID: \"0a768b62-44a4-4493-bd0d-2a536c55370d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.651111 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a768b62-44a4-4493-bd0d-2a536c55370d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm\" (UID: \"0a768b62-44a4-4493-bd0d-2a536c55370d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.651472 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a768b62-44a4-4493-bd0d-2a536c55370d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm\" (UID: \"0a768b62-44a4-4493-bd0d-2a536c55370d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.753510 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjmbb\" (UniqueName: \"kubernetes.io/projected/0a768b62-44a4-4493-bd0d-2a536c55370d-kube-api-access-kjmbb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm\" (UID: \"0a768b62-44a4-4493-bd0d-2a536c55370d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.753585 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a768b62-44a4-4493-bd0d-2a536c55370d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm\" (UID: \"0a768b62-44a4-4493-bd0d-2a536c55370d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.753696 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a768b62-44a4-4493-bd0d-2a536c55370d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm\" (UID: \"0a768b62-44a4-4493-bd0d-2a536c55370d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.758152 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a768b62-44a4-4493-bd0d-2a536c55370d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm\" (UID: \"0a768b62-44a4-4493-bd0d-2a536c55370d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.760103 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a768b62-44a4-4493-bd0d-2a536c55370d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm\" (UID: \"0a768b62-44a4-4493-bd0d-2a536c55370d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.773930 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjmbb\" (UniqueName: \"kubernetes.io/projected/0a768b62-44a4-4493-bd0d-2a536c55370d-kube-api-access-kjmbb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm\" (UID: \"0a768b62-44a4-4493-bd0d-2a536c55370d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:05:57 crc kubenswrapper[4857]: I1201 22:05:57.959909 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:05:58 crc kubenswrapper[4857]: I1201 22:05:58.323340 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm"] Dec 01 22:05:58 crc kubenswrapper[4857]: I1201 22:05:58.534775 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" event={"ID":"0a768b62-44a4-4493-bd0d-2a536c55370d","Type":"ContainerStarted","Data":"616db0e364f52755b2bc5ac880ba679fea235a82f18917e22b3ddea95cd23aaf"} Dec 01 22:05:58 crc kubenswrapper[4857]: I1201 22:05:58.834986 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:05:58 crc kubenswrapper[4857]: E1201 22:05:58.836069 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:05:59 crc kubenswrapper[4857]: I1201 22:05:59.546214 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" event={"ID":"0a768b62-44a4-4493-bd0d-2a536c55370d","Type":"ContainerStarted","Data":"284e7829a2df9e81ed44aa54e2f8277ab6435811f1128c2f528735d57e12df0f"} Dec 01 22:05:59 crc kubenswrapper[4857]: I1201 22:05:59.577501 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" podStartSLOduration=1.890653038 podStartE2EDuration="2.577462716s" podCreationTimestamp="2025-12-01 22:05:57 +0000 UTC" firstStartedPulling="2025-12-01 22:05:58.321900776 +0000 UTC m=+1876.811963093" lastFinishedPulling="2025-12-01 22:05:59.008710454 +0000 UTC m=+1877.498772771" observedRunningTime="2025-12-01 22:05:59.56914746 +0000 UTC m=+1878.059209777" watchObservedRunningTime="2025-12-01 22:05:59.577462716 +0000 UTC m=+1878.067525073" Dec 01 22:06:11 crc kubenswrapper[4857]: I1201 22:06:11.846393 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:06:11 crc kubenswrapper[4857]: E1201 22:06:11.847451 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:06:21 crc kubenswrapper[4857]: I1201 22:06:21.218991 4857 scope.go:117] "RemoveContainer" containerID="7e4657a3c7e681a255f82c8795900d576db2a774ab704fc3812ad8867925b769" Dec 01 22:06:21 crc kubenswrapper[4857]: I1201 22:06:21.319010 4857 scope.go:117] "RemoveContainer" containerID="6e264677551a6c313a6fa79f7941ae8b1ce5ea6b290d42d7ef24c21e8963a6d5" Dec 01 22:06:21 crc kubenswrapper[4857]: I1201 22:06:21.374452 4857 scope.go:117] "RemoveContainer" containerID="7d379a633511ffa997abc9a5932efaa8f962b1162a5507e79fe062fcd85a4ce1" Dec 01 22:06:21 crc kubenswrapper[4857]: I1201 22:06:21.423496 4857 scope.go:117] "RemoveContainer" containerID="a7330b1386f3883c5a5a6f43655ee634bc0e7398139f79dd64fc1c3c79b7bd42" Dec 01 22:06:21 crc kubenswrapper[4857]: I1201 22:06:21.451125 4857 scope.go:117] "RemoveContainer" containerID="dd26e7f56729f2b810b5b4d6c4062cbe94533fb3a1033c5c77df2624194787f1" Dec 01 22:06:21 crc kubenswrapper[4857]: I1201 22:06:21.497465 4857 scope.go:117] "RemoveContainer" containerID="abf1b62202614efcf7bfea038b20ef4221705a3afe75a406ca46b3134cca079a" Dec 01 22:06:24 crc kubenswrapper[4857]: I1201 22:06:24.835946 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:06:24 crc kubenswrapper[4857]: E1201 22:06:24.837247 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.058138 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-dhc4w"] Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.063621 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-dhc4w"] Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.072513 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-cb18-account-create-update-q2fbk"] Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.085111 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-d8e3-account-create-update-nrctx"] Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.091176 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-7c37-account-create-update-rzsv8"] Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.120111 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-cb18-account-create-update-q2fbk"] Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.161109 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-d8e3-account-create-update-nrctx"] Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.183096 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-7c37-account-create-update-rzsv8"] Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.291469 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-ktkxd"] Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.317098 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-ktkxd"] Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.329711 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-cjk48"] Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.342921 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-cjk48"] Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.850808 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16aaf417-9ea4-4f49-ad7b-77064d9fcb8c" path="/var/lib/kubelet/pods/16aaf417-9ea4-4f49-ad7b-77064d9fcb8c/volumes" Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.853593 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="574a17c8-2c54-41d2-8390-aff1f2220638" path="/var/lib/kubelet/pods/574a17c8-2c54-41d2-8390-aff1f2220638/volumes" Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.855163 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9638968d-42b2-46ab-b02a-72d044325fa9" path="/var/lib/kubelet/pods/9638968d-42b2-46ab-b02a-72d044325fa9/volumes" Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.856354 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7174f1f-8255-4dfe-a230-5128ba63571e" path="/var/lib/kubelet/pods/a7174f1f-8255-4dfe-a230-5128ba63571e/volumes" Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.858735 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc85532c-b753-4a79-ad36-621f558d03d0" path="/var/lib/kubelet/pods/cc85532c-b753-4a79-ad36-621f558d03d0/volumes" Dec 01 22:06:27 crc kubenswrapper[4857]: I1201 22:06:27.860132 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9e73cb1-7588-46b5-9014-efe779354e1d" path="/var/lib/kubelet/pods/d9e73cb1-7588-46b5-9014-efe779354e1d/volumes" Dec 01 22:06:36 crc kubenswrapper[4857]: I1201 22:06:36.836092 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:06:36 crc kubenswrapper[4857]: E1201 22:06:36.837198 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:06:49 crc kubenswrapper[4857]: I1201 22:06:49.835498 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:06:49 crc kubenswrapper[4857]: E1201 22:06:49.836536 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:07:02 crc kubenswrapper[4857]: I1201 22:07:02.055876 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-r7zf9"] Dec 01 22:07:02 crc kubenswrapper[4857]: I1201 22:07:02.073329 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-r7zf9"] Dec 01 22:07:02 crc kubenswrapper[4857]: I1201 22:07:02.835449 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:07:02 crc kubenswrapper[4857]: E1201 22:07:02.835710 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:07:03 crc kubenswrapper[4857]: I1201 22:07:03.851952 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0906fcdc-a402-46cf-bf71-fcd7543a9034" path="/var/lib/kubelet/pods/0906fcdc-a402-46cf-bf71-fcd7543a9034/volumes" Dec 01 22:07:14 crc kubenswrapper[4857]: I1201 22:07:14.835500 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:07:14 crc kubenswrapper[4857]: E1201 22:07:14.836395 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:07:17 crc kubenswrapper[4857]: I1201 22:07:17.535145 4857 generic.go:334] "Generic (PLEG): container finished" podID="0a768b62-44a4-4493-bd0d-2a536c55370d" containerID="284e7829a2df9e81ed44aa54e2f8277ab6435811f1128c2f528735d57e12df0f" exitCode=0 Dec 01 22:07:17 crc kubenswrapper[4857]: I1201 22:07:17.535199 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" event={"ID":"0a768b62-44a4-4493-bd0d-2a536c55370d","Type":"ContainerDied","Data":"284e7829a2df9e81ed44aa54e2f8277ab6435811f1128c2f528735d57e12df0f"} Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.067081 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.170506 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a768b62-44a4-4493-bd0d-2a536c55370d-inventory\") pod \"0a768b62-44a4-4493-bd0d-2a536c55370d\" (UID: \"0a768b62-44a4-4493-bd0d-2a536c55370d\") " Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.170594 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjmbb\" (UniqueName: \"kubernetes.io/projected/0a768b62-44a4-4493-bd0d-2a536c55370d-kube-api-access-kjmbb\") pod \"0a768b62-44a4-4493-bd0d-2a536c55370d\" (UID: \"0a768b62-44a4-4493-bd0d-2a536c55370d\") " Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.170658 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a768b62-44a4-4493-bd0d-2a536c55370d-ssh-key\") pod \"0a768b62-44a4-4493-bd0d-2a536c55370d\" (UID: \"0a768b62-44a4-4493-bd0d-2a536c55370d\") " Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.179588 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a768b62-44a4-4493-bd0d-2a536c55370d-kube-api-access-kjmbb" (OuterVolumeSpecName: "kube-api-access-kjmbb") pod "0a768b62-44a4-4493-bd0d-2a536c55370d" (UID: "0a768b62-44a4-4493-bd0d-2a536c55370d"). InnerVolumeSpecName "kube-api-access-kjmbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.226238 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a768b62-44a4-4493-bd0d-2a536c55370d-inventory" (OuterVolumeSpecName: "inventory") pod "0a768b62-44a4-4493-bd0d-2a536c55370d" (UID: "0a768b62-44a4-4493-bd0d-2a536c55370d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.228278 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a768b62-44a4-4493-bd0d-2a536c55370d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0a768b62-44a4-4493-bd0d-2a536c55370d" (UID: "0a768b62-44a4-4493-bd0d-2a536c55370d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.274742 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a768b62-44a4-4493-bd0d-2a536c55370d-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.274797 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjmbb\" (UniqueName: \"kubernetes.io/projected/0a768b62-44a4-4493-bd0d-2a536c55370d-kube-api-access-kjmbb\") on node \"crc\" DevicePath \"\"" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.274836 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a768b62-44a4-4493-bd0d-2a536c55370d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.558054 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" event={"ID":"0a768b62-44a4-4493-bd0d-2a536c55370d","Type":"ContainerDied","Data":"616db0e364f52755b2bc5ac880ba679fea235a82f18917e22b3ddea95cd23aaf"} Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.558305 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="616db0e364f52755b2bc5ac880ba679fea235a82f18917e22b3ddea95cd23aaf" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.558189 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.716469 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9"] Dec 01 22:07:19 crc kubenswrapper[4857]: E1201 22:07:19.716922 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a768b62-44a4-4493-bd0d-2a536c55370d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.716946 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a768b62-44a4-4493-bd0d-2a536c55370d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.718149 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a768b62-44a4-4493-bd0d-2a536c55370d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.718940 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.723211 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.723447 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.723537 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.723650 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.740903 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9"] Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.916578 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e74258df-2d15-4928-993f-fac533426f97-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9\" (UID: \"e74258df-2d15-4928-993f-fac533426f97\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.916672 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e74258df-2d15-4928-993f-fac533426f97-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9\" (UID: \"e74258df-2d15-4928-993f-fac533426f97\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:19 crc kubenswrapper[4857]: I1201 22:07:19.916726 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v26pm\" (UniqueName: \"kubernetes.io/projected/e74258df-2d15-4928-993f-fac533426f97-kube-api-access-v26pm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9\" (UID: \"e74258df-2d15-4928-993f-fac533426f97\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:20 crc kubenswrapper[4857]: I1201 22:07:20.019051 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e74258df-2d15-4928-993f-fac533426f97-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9\" (UID: \"e74258df-2d15-4928-993f-fac533426f97\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:20 crc kubenswrapper[4857]: I1201 22:07:20.019143 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e74258df-2d15-4928-993f-fac533426f97-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9\" (UID: \"e74258df-2d15-4928-993f-fac533426f97\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:20 crc kubenswrapper[4857]: I1201 22:07:20.019182 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v26pm\" (UniqueName: \"kubernetes.io/projected/e74258df-2d15-4928-993f-fac533426f97-kube-api-access-v26pm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9\" (UID: \"e74258df-2d15-4928-993f-fac533426f97\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:20 crc kubenswrapper[4857]: I1201 22:07:20.025999 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e74258df-2d15-4928-993f-fac533426f97-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9\" (UID: \"e74258df-2d15-4928-993f-fac533426f97\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:20 crc kubenswrapper[4857]: I1201 22:07:20.027751 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e74258df-2d15-4928-993f-fac533426f97-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9\" (UID: \"e74258df-2d15-4928-993f-fac533426f97\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:20 crc kubenswrapper[4857]: I1201 22:07:20.043968 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v26pm\" (UniqueName: \"kubernetes.io/projected/e74258df-2d15-4928-993f-fac533426f97-kube-api-access-v26pm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9\" (UID: \"e74258df-2d15-4928-993f-fac533426f97\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:20 crc kubenswrapper[4857]: I1201 22:07:20.047264 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:20 crc kubenswrapper[4857]: I1201 22:07:20.707075 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9"] Dec 01 22:07:21 crc kubenswrapper[4857]: I1201 22:07:21.587707 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" event={"ID":"e74258df-2d15-4928-993f-fac533426f97","Type":"ContainerStarted","Data":"35de9f1b6b3b74750b43dcfc9a19435f4233370f7c2b2d2b7b20d2fb42ea0c78"} Dec 01 22:07:21 crc kubenswrapper[4857]: I1201 22:07:21.588531 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" event={"ID":"e74258df-2d15-4928-993f-fac533426f97","Type":"ContainerStarted","Data":"3a3f1f4419f1b55ebab6d13ca5179158b8e1656f01ec1b1376f227e1fa98b75c"} Dec 01 22:07:21 crc kubenswrapper[4857]: I1201 22:07:21.624859 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" podStartSLOduration=2.029141702 podStartE2EDuration="2.62474146s" podCreationTimestamp="2025-12-01 22:07:19 +0000 UTC" firstStartedPulling="2025-12-01 22:07:20.7129433 +0000 UTC m=+1959.203005627" lastFinishedPulling="2025-12-01 22:07:21.308543058 +0000 UTC m=+1959.798605385" observedRunningTime="2025-12-01 22:07:21.611948728 +0000 UTC m=+1960.102011125" watchObservedRunningTime="2025-12-01 22:07:21.62474146 +0000 UTC m=+1960.114803807" Dec 01 22:07:21 crc kubenswrapper[4857]: I1201 22:07:21.670149 4857 scope.go:117] "RemoveContainer" containerID="fd5e469f0ee3459dff1e28e46ccefe6cbfd8604cbb8469fe8a30030c53c8c130" Dec 01 22:07:21 crc kubenswrapper[4857]: I1201 22:07:21.691484 4857 scope.go:117] "RemoveContainer" containerID="9e3539f6bb3a98cd815f508138d9ad0a369f92eba591b937717647a0ca512a72" Dec 01 22:07:21 crc kubenswrapper[4857]: I1201 22:07:21.779489 4857 scope.go:117] "RemoveContainer" containerID="78ff7551210d68af64ef40de04477bb2d7bd350b72c7497937d3602bc511379e" Dec 01 22:07:21 crc kubenswrapper[4857]: I1201 22:07:21.818696 4857 scope.go:117] "RemoveContainer" containerID="f68ec44666a1cf9a01d3df410f654e5967c439d9ffc2cbdeb314b5120496b0e1" Dec 01 22:07:21 crc kubenswrapper[4857]: I1201 22:07:21.840484 4857 scope.go:117] "RemoveContainer" containerID="3d5c157e719b635007d763a137b271d169fb1abf27dd50770eb6d1b582825668" Dec 01 22:07:21 crc kubenswrapper[4857]: I1201 22:07:21.861308 4857 scope.go:117] "RemoveContainer" containerID="ad4dfe871a26d8c5d1ed8e6808025f3b4fc406625ed549bb498fb4d8cae5c9e9" Dec 01 22:07:21 crc kubenswrapper[4857]: I1201 22:07:21.886409 4857 scope.go:117] "RemoveContainer" containerID="dbb202fdc130888f71b8909accc0b79a3ea00018f73fdf434eee22ca8109ceb4" Dec 01 22:07:25 crc kubenswrapper[4857]: I1201 22:07:25.034521 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-m4rd7"] Dec 01 22:07:25 crc kubenswrapper[4857]: I1201 22:07:25.041807 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-m4rd7"] Dec 01 22:07:25 crc kubenswrapper[4857]: I1201 22:07:25.853577 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70f52857-7d57-421e-8bf0-b849a8c02ffe" path="/var/lib/kubelet/pods/70f52857-7d57-421e-8bf0-b849a8c02ffe/volumes" Dec 01 22:07:26 crc kubenswrapper[4857]: I1201 22:07:26.836767 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:07:26 crc kubenswrapper[4857]: E1201 22:07:26.838133 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:07:27 crc kubenswrapper[4857]: I1201 22:07:27.660436 4857 generic.go:334] "Generic (PLEG): container finished" podID="e74258df-2d15-4928-993f-fac533426f97" containerID="35de9f1b6b3b74750b43dcfc9a19435f4233370f7c2b2d2b7b20d2fb42ea0c78" exitCode=0 Dec 01 22:07:27 crc kubenswrapper[4857]: I1201 22:07:27.660501 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" event={"ID":"e74258df-2d15-4928-993f-fac533426f97","Type":"ContainerDied","Data":"35de9f1b6b3b74750b43dcfc9a19435f4233370f7c2b2d2b7b20d2fb42ea0c78"} Dec 01 22:07:28 crc kubenswrapper[4857]: I1201 22:07:28.063444 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-v9qs5"] Dec 01 22:07:28 crc kubenswrapper[4857]: I1201 22:07:28.074947 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-v9qs5"] Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.205119 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.328343 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e74258df-2d15-4928-993f-fac533426f97-inventory\") pod \"e74258df-2d15-4928-993f-fac533426f97\" (UID: \"e74258df-2d15-4928-993f-fac533426f97\") " Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.328431 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e74258df-2d15-4928-993f-fac533426f97-ssh-key\") pod \"e74258df-2d15-4928-993f-fac533426f97\" (UID: \"e74258df-2d15-4928-993f-fac533426f97\") " Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.328488 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v26pm\" (UniqueName: \"kubernetes.io/projected/e74258df-2d15-4928-993f-fac533426f97-kube-api-access-v26pm\") pod \"e74258df-2d15-4928-993f-fac533426f97\" (UID: \"e74258df-2d15-4928-993f-fac533426f97\") " Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.338381 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e74258df-2d15-4928-993f-fac533426f97-kube-api-access-v26pm" (OuterVolumeSpecName: "kube-api-access-v26pm") pod "e74258df-2d15-4928-993f-fac533426f97" (UID: "e74258df-2d15-4928-993f-fac533426f97"). InnerVolumeSpecName "kube-api-access-v26pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.366305 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74258df-2d15-4928-993f-fac533426f97-inventory" (OuterVolumeSpecName: "inventory") pod "e74258df-2d15-4928-993f-fac533426f97" (UID: "e74258df-2d15-4928-993f-fac533426f97"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.385928 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74258df-2d15-4928-993f-fac533426f97-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e74258df-2d15-4928-993f-fac533426f97" (UID: "e74258df-2d15-4928-993f-fac533426f97"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.432887 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e74258df-2d15-4928-993f-fac533426f97-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.432998 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e74258df-2d15-4928-993f-fac533426f97-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.433020 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v26pm\" (UniqueName: \"kubernetes.io/projected/e74258df-2d15-4928-993f-fac533426f97-kube-api-access-v26pm\") on node \"crc\" DevicePath \"\"" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.684599 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" event={"ID":"e74258df-2d15-4928-993f-fac533426f97","Type":"ContainerDied","Data":"3a3f1f4419f1b55ebab6d13ca5179158b8e1656f01ec1b1376f227e1fa98b75c"} Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.684655 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a3f1f4419f1b55ebab6d13ca5179158b8e1656f01ec1b1376f227e1fa98b75c" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.684689 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.780355 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx"] Dec 01 22:07:29 crc kubenswrapper[4857]: E1201 22:07:29.781201 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74258df-2d15-4928-993f-fac533426f97" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.781244 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74258df-2d15-4928-993f-fac533426f97" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.781768 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e74258df-2d15-4928-993f-fac533426f97" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.783284 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.786776 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.788657 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.789334 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.789607 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.796467 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx"] Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.841082 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/696c86fa-0abd-4bae-9821-f570a70ad6e0-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2zzdx\" (UID: \"696c86fa-0abd-4bae-9821-f570a70ad6e0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.841167 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/696c86fa-0abd-4bae-9821-f570a70ad6e0-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2zzdx\" (UID: \"696c86fa-0abd-4bae-9821-f570a70ad6e0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.841205 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8f97\" (UniqueName: \"kubernetes.io/projected/696c86fa-0abd-4bae-9821-f570a70ad6e0-kube-api-access-l8f97\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2zzdx\" (UID: \"696c86fa-0abd-4bae-9821-f570a70ad6e0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.848324 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9df6644a-8870-44de-8e78-51f76e062f85" path="/var/lib/kubelet/pods/9df6644a-8870-44de-8e78-51f76e062f85/volumes" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.943233 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/696c86fa-0abd-4bae-9821-f570a70ad6e0-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2zzdx\" (UID: \"696c86fa-0abd-4bae-9821-f570a70ad6e0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.943312 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8f97\" (UniqueName: \"kubernetes.io/projected/696c86fa-0abd-4bae-9821-f570a70ad6e0-kube-api-access-l8f97\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2zzdx\" (UID: \"696c86fa-0abd-4bae-9821-f570a70ad6e0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.943653 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/696c86fa-0abd-4bae-9821-f570a70ad6e0-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2zzdx\" (UID: \"696c86fa-0abd-4bae-9821-f570a70ad6e0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.949247 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/696c86fa-0abd-4bae-9821-f570a70ad6e0-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2zzdx\" (UID: \"696c86fa-0abd-4bae-9821-f570a70ad6e0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.949547 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/696c86fa-0abd-4bae-9821-f570a70ad6e0-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2zzdx\" (UID: \"696c86fa-0abd-4bae-9821-f570a70ad6e0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:07:29 crc kubenswrapper[4857]: I1201 22:07:29.975321 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8f97\" (UniqueName: \"kubernetes.io/projected/696c86fa-0abd-4bae-9821-f570a70ad6e0-kube-api-access-l8f97\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2zzdx\" (UID: \"696c86fa-0abd-4bae-9821-f570a70ad6e0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:07:30 crc kubenswrapper[4857]: I1201 22:07:30.110107 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:07:30 crc kubenswrapper[4857]: I1201 22:07:30.511556 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx"] Dec 01 22:07:30 crc kubenswrapper[4857]: I1201 22:07:30.693249 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" event={"ID":"696c86fa-0abd-4bae-9821-f570a70ad6e0","Type":"ContainerStarted","Data":"75c08903c8a140b3d1d7109800914a0a963c553259d364e28e78f08ae6c97fa8"} Dec 01 22:07:31 crc kubenswrapper[4857]: I1201 22:07:31.702516 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" event={"ID":"696c86fa-0abd-4bae-9821-f570a70ad6e0","Type":"ContainerStarted","Data":"b3ff50109b0b470234984f20b480ea3ae17678476f9c2b1e2c7b36addbd8e581"} Dec 01 22:07:31 crc kubenswrapper[4857]: I1201 22:07:31.729189 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" podStartSLOduration=1.91887623 podStartE2EDuration="2.729167535s" podCreationTimestamp="2025-12-01 22:07:29 +0000 UTC" firstStartedPulling="2025-12-01 22:07:30.510189382 +0000 UTC m=+1969.000251699" lastFinishedPulling="2025-12-01 22:07:31.320480657 +0000 UTC m=+1969.810543004" observedRunningTime="2025-12-01 22:07:31.72239735 +0000 UTC m=+1970.212459697" watchObservedRunningTime="2025-12-01 22:07:31.729167535 +0000 UTC m=+1970.219229852" Dec 01 22:07:38 crc kubenswrapper[4857]: I1201 22:07:38.835962 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:07:38 crc kubenswrapper[4857]: E1201 22:07:38.837474 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:07:52 crc kubenswrapper[4857]: I1201 22:07:52.835791 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:07:52 crc kubenswrapper[4857]: E1201 22:07:52.837199 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:08:07 crc kubenswrapper[4857]: I1201 22:08:07.835550 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:08:08 crc kubenswrapper[4857]: I1201 22:08:08.141786 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"d86df5e45e1b4e70e1cb13ada553f4b40c37fb697e38ddf8687ffc69d2758e7f"} Dec 01 22:08:12 crc kubenswrapper[4857]: I1201 22:08:12.059095 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-lwhqc"] Dec 01 22:08:12 crc kubenswrapper[4857]: I1201 22:08:12.069674 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-lwhqc"] Dec 01 22:08:13 crc kubenswrapper[4857]: I1201 22:08:13.856442 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="345660e3-c8c5-45ef-a5e5-315d7254a41c" path="/var/lib/kubelet/pods/345660e3-c8c5-45ef-a5e5-315d7254a41c/volumes" Dec 01 22:08:15 crc kubenswrapper[4857]: E1201 22:08:15.827691 4857 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod696c86fa_0abd_4bae_9821_f570a70ad6e0.slice/crio-b3ff50109b0b470234984f20b480ea3ae17678476f9c2b1e2c7b36addbd8e581.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod696c86fa_0abd_4bae_9821_f570a70ad6e0.slice/crio-conmon-b3ff50109b0b470234984f20b480ea3ae17678476f9c2b1e2c7b36addbd8e581.scope\": RecentStats: unable to find data in memory cache]" Dec 01 22:08:16 crc kubenswrapper[4857]: I1201 22:08:16.230283 4857 generic.go:334] "Generic (PLEG): container finished" podID="696c86fa-0abd-4bae-9821-f570a70ad6e0" containerID="b3ff50109b0b470234984f20b480ea3ae17678476f9c2b1e2c7b36addbd8e581" exitCode=0 Dec 01 22:08:16 crc kubenswrapper[4857]: I1201 22:08:16.230328 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" event={"ID":"696c86fa-0abd-4bae-9821-f570a70ad6e0","Type":"ContainerDied","Data":"b3ff50109b0b470234984f20b480ea3ae17678476f9c2b1e2c7b36addbd8e581"} Dec 01 22:08:17 crc kubenswrapper[4857]: I1201 22:08:17.750947 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:08:17 crc kubenswrapper[4857]: I1201 22:08:17.872479 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/696c86fa-0abd-4bae-9821-f570a70ad6e0-ssh-key\") pod \"696c86fa-0abd-4bae-9821-f570a70ad6e0\" (UID: \"696c86fa-0abd-4bae-9821-f570a70ad6e0\") " Dec 01 22:08:17 crc kubenswrapper[4857]: I1201 22:08:17.872637 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8f97\" (UniqueName: \"kubernetes.io/projected/696c86fa-0abd-4bae-9821-f570a70ad6e0-kube-api-access-l8f97\") pod \"696c86fa-0abd-4bae-9821-f570a70ad6e0\" (UID: \"696c86fa-0abd-4bae-9821-f570a70ad6e0\") " Dec 01 22:08:17 crc kubenswrapper[4857]: I1201 22:08:17.872933 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/696c86fa-0abd-4bae-9821-f570a70ad6e0-inventory\") pod \"696c86fa-0abd-4bae-9821-f570a70ad6e0\" (UID: \"696c86fa-0abd-4bae-9821-f570a70ad6e0\") " Dec 01 22:08:17 crc kubenswrapper[4857]: I1201 22:08:17.883232 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/696c86fa-0abd-4bae-9821-f570a70ad6e0-kube-api-access-l8f97" (OuterVolumeSpecName: "kube-api-access-l8f97") pod "696c86fa-0abd-4bae-9821-f570a70ad6e0" (UID: "696c86fa-0abd-4bae-9821-f570a70ad6e0"). InnerVolumeSpecName "kube-api-access-l8f97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:08:17 crc kubenswrapper[4857]: I1201 22:08:17.912280 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/696c86fa-0abd-4bae-9821-f570a70ad6e0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "696c86fa-0abd-4bae-9821-f570a70ad6e0" (UID: "696c86fa-0abd-4bae-9821-f570a70ad6e0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:08:17 crc kubenswrapper[4857]: I1201 22:08:17.930158 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/696c86fa-0abd-4bae-9821-f570a70ad6e0-inventory" (OuterVolumeSpecName: "inventory") pod "696c86fa-0abd-4bae-9821-f570a70ad6e0" (UID: "696c86fa-0abd-4bae-9821-f570a70ad6e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:08:17 crc kubenswrapper[4857]: I1201 22:08:17.975002 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/696c86fa-0abd-4bae-9821-f570a70ad6e0-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:08:17 crc kubenswrapper[4857]: I1201 22:08:17.975028 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/696c86fa-0abd-4bae-9821-f570a70ad6e0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:08:17 crc kubenswrapper[4857]: I1201 22:08:17.975055 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8f97\" (UniqueName: \"kubernetes.io/projected/696c86fa-0abd-4bae-9821-f570a70ad6e0-kube-api-access-l8f97\") on node \"crc\" DevicePath \"\"" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.265058 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" event={"ID":"696c86fa-0abd-4bae-9821-f570a70ad6e0","Type":"ContainerDied","Data":"75c08903c8a140b3d1d7109800914a0a963c553259d364e28e78f08ae6c97fa8"} Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.265103 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75c08903c8a140b3d1d7109800914a0a963c553259d364e28e78f08ae6c97fa8" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.265132 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2zzdx" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.368176 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz"] Dec 01 22:08:18 crc kubenswrapper[4857]: E1201 22:08:18.369026 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="696c86fa-0abd-4bae-9821-f570a70ad6e0" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.369076 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="696c86fa-0abd-4bae-9821-f570a70ad6e0" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.370166 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="696c86fa-0abd-4bae-9821-f570a70ad6e0" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.371358 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.375576 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.375832 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.375933 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.375936 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.377600 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz"] Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.492201 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3196a47-b0ed-4f5d-a639-496ff78d1acd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7mccz\" (UID: \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.492269 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csgr8\" (UniqueName: \"kubernetes.io/projected/e3196a47-b0ed-4f5d-a639-496ff78d1acd-kube-api-access-csgr8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7mccz\" (UID: \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.492329 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3196a47-b0ed-4f5d-a639-496ff78d1acd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7mccz\" (UID: \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.595472 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3196a47-b0ed-4f5d-a639-496ff78d1acd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7mccz\" (UID: \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.595888 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csgr8\" (UniqueName: \"kubernetes.io/projected/e3196a47-b0ed-4f5d-a639-496ff78d1acd-kube-api-access-csgr8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7mccz\" (UID: \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.596160 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3196a47-b0ed-4f5d-a639-496ff78d1acd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7mccz\" (UID: \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.601226 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3196a47-b0ed-4f5d-a639-496ff78d1acd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7mccz\" (UID: \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.601378 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3196a47-b0ed-4f5d-a639-496ff78d1acd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7mccz\" (UID: \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.613718 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csgr8\" (UniqueName: \"kubernetes.io/projected/e3196a47-b0ed-4f5d-a639-496ff78d1acd-kube-api-access-csgr8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7mccz\" (UID: \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:08:18 crc kubenswrapper[4857]: I1201 22:08:18.696672 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:08:19 crc kubenswrapper[4857]: I1201 22:08:19.064132 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz"] Dec 01 22:08:19 crc kubenswrapper[4857]: I1201 22:08:19.278329 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" event={"ID":"e3196a47-b0ed-4f5d-a639-496ff78d1acd","Type":"ContainerStarted","Data":"2a977fe7e8ec3cc637226845aae8e9a59ac39aebc0671ade8d45d862df5bfe38"} Dec 01 22:08:20 crc kubenswrapper[4857]: I1201 22:08:20.291536 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" event={"ID":"e3196a47-b0ed-4f5d-a639-496ff78d1acd","Type":"ContainerStarted","Data":"fdeb887b128fadfaaa9595b9e7060ea10bb5e0b77cf52a5e87510c9071bcb3f9"} Dec 01 22:08:20 crc kubenswrapper[4857]: I1201 22:08:20.325361 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" podStartSLOduration=1.751881684 podStartE2EDuration="2.325328993s" podCreationTimestamp="2025-12-01 22:08:18 +0000 UTC" firstStartedPulling="2025-12-01 22:08:19.066307126 +0000 UTC m=+2017.556369443" lastFinishedPulling="2025-12-01 22:08:19.639754405 +0000 UTC m=+2018.129816752" observedRunningTime="2025-12-01 22:08:20.312027525 +0000 UTC m=+2018.802089872" watchObservedRunningTime="2025-12-01 22:08:20.325328993 +0000 UTC m=+2018.815391350" Dec 01 22:08:22 crc kubenswrapper[4857]: I1201 22:08:22.066711 4857 scope.go:117] "RemoveContainer" containerID="d78b0cfe9bd8d514b110c57c6cac5a8599248656c2b26a32beb8199cf1eae8c6" Dec 01 22:08:22 crc kubenswrapper[4857]: I1201 22:08:22.148309 4857 scope.go:117] "RemoveContainer" containerID="180c474545583e5dd7bd2a9be4851eaa2bccaf82f0383d261a55dedee7589fad" Dec 01 22:08:22 crc kubenswrapper[4857]: I1201 22:08:22.209856 4857 scope.go:117] "RemoveContainer" containerID="f64fce28262cd259de4d3107ccf2fd6b7a1b3c2c8b4babaae8b518b5223d8aba" Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.020105 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9d749"] Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.028508 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.039035 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9d749"] Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.166857 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w2tn\" (UniqueName: \"kubernetes.io/projected/51b8e39d-30ef-4325-91ec-8b46219e093c-kube-api-access-4w2tn\") pod \"redhat-operators-9d749\" (UID: \"51b8e39d-30ef-4325-91ec-8b46219e093c\") " pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.166921 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51b8e39d-30ef-4325-91ec-8b46219e093c-utilities\") pod \"redhat-operators-9d749\" (UID: \"51b8e39d-30ef-4325-91ec-8b46219e093c\") " pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.167016 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51b8e39d-30ef-4325-91ec-8b46219e093c-catalog-content\") pod \"redhat-operators-9d749\" (UID: \"51b8e39d-30ef-4325-91ec-8b46219e093c\") " pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.269313 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w2tn\" (UniqueName: \"kubernetes.io/projected/51b8e39d-30ef-4325-91ec-8b46219e093c-kube-api-access-4w2tn\") pod \"redhat-operators-9d749\" (UID: \"51b8e39d-30ef-4325-91ec-8b46219e093c\") " pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.269382 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51b8e39d-30ef-4325-91ec-8b46219e093c-utilities\") pod \"redhat-operators-9d749\" (UID: \"51b8e39d-30ef-4325-91ec-8b46219e093c\") " pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.269502 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51b8e39d-30ef-4325-91ec-8b46219e093c-catalog-content\") pod \"redhat-operators-9d749\" (UID: \"51b8e39d-30ef-4325-91ec-8b46219e093c\") " pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.270151 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51b8e39d-30ef-4325-91ec-8b46219e093c-catalog-content\") pod \"redhat-operators-9d749\" (UID: \"51b8e39d-30ef-4325-91ec-8b46219e093c\") " pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.270322 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51b8e39d-30ef-4325-91ec-8b46219e093c-utilities\") pod \"redhat-operators-9d749\" (UID: \"51b8e39d-30ef-4325-91ec-8b46219e093c\") " pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.296743 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w2tn\" (UniqueName: \"kubernetes.io/projected/51b8e39d-30ef-4325-91ec-8b46219e093c-kube-api-access-4w2tn\") pod \"redhat-operators-9d749\" (UID: \"51b8e39d-30ef-4325-91ec-8b46219e093c\") " pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.369183 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:30 crc kubenswrapper[4857]: I1201 22:08:30.868232 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9d749"] Dec 01 22:08:30 crc kubenswrapper[4857]: W1201 22:08:30.879885 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51b8e39d_30ef_4325_91ec_8b46219e093c.slice/crio-d6262d23e07e04bd7b2d55af127016e9cd5d355cc30c2f28b3012502078248de WatchSource:0}: Error finding container d6262d23e07e04bd7b2d55af127016e9cd5d355cc30c2f28b3012502078248de: Status 404 returned error can't find the container with id d6262d23e07e04bd7b2d55af127016e9cd5d355cc30c2f28b3012502078248de Dec 01 22:08:31 crc kubenswrapper[4857]: I1201 22:08:31.462087 4857 generic.go:334] "Generic (PLEG): container finished" podID="51b8e39d-30ef-4325-91ec-8b46219e093c" containerID="43d6c1097a6cb03e1d0d353c1d396ff75759e65e2c1d8dd993636b7210c4ce0b" exitCode=0 Dec 01 22:08:31 crc kubenswrapper[4857]: I1201 22:08:31.462201 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9d749" event={"ID":"51b8e39d-30ef-4325-91ec-8b46219e093c","Type":"ContainerDied","Data":"43d6c1097a6cb03e1d0d353c1d396ff75759e65e2c1d8dd993636b7210c4ce0b"} Dec 01 22:08:31 crc kubenswrapper[4857]: I1201 22:08:31.462374 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9d749" event={"ID":"51b8e39d-30ef-4325-91ec-8b46219e093c","Type":"ContainerStarted","Data":"d6262d23e07e04bd7b2d55af127016e9cd5d355cc30c2f28b3012502078248de"} Dec 01 22:08:33 crc kubenswrapper[4857]: I1201 22:08:33.497919 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9d749" event={"ID":"51b8e39d-30ef-4325-91ec-8b46219e093c","Type":"ContainerStarted","Data":"d74e2032540592afd088f7fa9601de559346100acee448f812a6080f4b577b85"} Dec 01 22:08:34 crc kubenswrapper[4857]: I1201 22:08:34.515179 4857 generic.go:334] "Generic (PLEG): container finished" podID="51b8e39d-30ef-4325-91ec-8b46219e093c" containerID="d74e2032540592afd088f7fa9601de559346100acee448f812a6080f4b577b85" exitCode=0 Dec 01 22:08:34 crc kubenswrapper[4857]: I1201 22:08:34.515240 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9d749" event={"ID":"51b8e39d-30ef-4325-91ec-8b46219e093c","Type":"ContainerDied","Data":"d74e2032540592afd088f7fa9601de559346100acee448f812a6080f4b577b85"} Dec 01 22:08:36 crc kubenswrapper[4857]: I1201 22:08:36.578848 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9d749" event={"ID":"51b8e39d-30ef-4325-91ec-8b46219e093c","Type":"ContainerStarted","Data":"96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8"} Dec 01 22:08:36 crc kubenswrapper[4857]: I1201 22:08:36.608777 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9d749" podStartSLOduration=4.017537548 podStartE2EDuration="7.608743216s" podCreationTimestamp="2025-12-01 22:08:29 +0000 UTC" firstStartedPulling="2025-12-01 22:08:31.463540243 +0000 UTC m=+2029.953602560" lastFinishedPulling="2025-12-01 22:08:35.054745901 +0000 UTC m=+2033.544808228" observedRunningTime="2025-12-01 22:08:36.605432754 +0000 UTC m=+2035.095495121" watchObservedRunningTime="2025-12-01 22:08:36.608743216 +0000 UTC m=+2035.098805573" Dec 01 22:08:40 crc kubenswrapper[4857]: I1201 22:08:40.370096 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:40 crc kubenswrapper[4857]: I1201 22:08:40.373229 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:41 crc kubenswrapper[4857]: I1201 22:08:41.462920 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9d749" podUID="51b8e39d-30ef-4325-91ec-8b46219e093c" containerName="registry-server" probeResult="failure" output=< Dec 01 22:08:41 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Dec 01 22:08:41 crc kubenswrapper[4857]: > Dec 01 22:08:50 crc kubenswrapper[4857]: I1201 22:08:50.457783 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:50 crc kubenswrapper[4857]: I1201 22:08:50.516667 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:50 crc kubenswrapper[4857]: I1201 22:08:50.714897 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9d749"] Dec 01 22:08:51 crc kubenswrapper[4857]: I1201 22:08:51.766400 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9d749" podUID="51b8e39d-30ef-4325-91ec-8b46219e093c" containerName="registry-server" containerID="cri-o://96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8" gracePeriod=2 Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.328187 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.434560 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w2tn\" (UniqueName: \"kubernetes.io/projected/51b8e39d-30ef-4325-91ec-8b46219e093c-kube-api-access-4w2tn\") pod \"51b8e39d-30ef-4325-91ec-8b46219e093c\" (UID: \"51b8e39d-30ef-4325-91ec-8b46219e093c\") " Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.434664 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51b8e39d-30ef-4325-91ec-8b46219e093c-catalog-content\") pod \"51b8e39d-30ef-4325-91ec-8b46219e093c\" (UID: \"51b8e39d-30ef-4325-91ec-8b46219e093c\") " Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.434777 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51b8e39d-30ef-4325-91ec-8b46219e093c-utilities\") pod \"51b8e39d-30ef-4325-91ec-8b46219e093c\" (UID: \"51b8e39d-30ef-4325-91ec-8b46219e093c\") " Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.436816 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51b8e39d-30ef-4325-91ec-8b46219e093c-utilities" (OuterVolumeSpecName: "utilities") pod "51b8e39d-30ef-4325-91ec-8b46219e093c" (UID: "51b8e39d-30ef-4325-91ec-8b46219e093c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.448074 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51b8e39d-30ef-4325-91ec-8b46219e093c-kube-api-access-4w2tn" (OuterVolumeSpecName: "kube-api-access-4w2tn") pod "51b8e39d-30ef-4325-91ec-8b46219e093c" (UID: "51b8e39d-30ef-4325-91ec-8b46219e093c"). InnerVolumeSpecName "kube-api-access-4w2tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.537301 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w2tn\" (UniqueName: \"kubernetes.io/projected/51b8e39d-30ef-4325-91ec-8b46219e093c-kube-api-access-4w2tn\") on node \"crc\" DevicePath \"\"" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.537333 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51b8e39d-30ef-4325-91ec-8b46219e093c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.573648 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51b8e39d-30ef-4325-91ec-8b46219e093c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51b8e39d-30ef-4325-91ec-8b46219e093c" (UID: "51b8e39d-30ef-4325-91ec-8b46219e093c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.639867 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51b8e39d-30ef-4325-91ec-8b46219e093c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.784639 4857 generic.go:334] "Generic (PLEG): container finished" podID="51b8e39d-30ef-4325-91ec-8b46219e093c" containerID="96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8" exitCode=0 Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.784703 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9d749" event={"ID":"51b8e39d-30ef-4325-91ec-8b46219e093c","Type":"ContainerDied","Data":"96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8"} Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.784722 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9d749" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.784755 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9d749" event={"ID":"51b8e39d-30ef-4325-91ec-8b46219e093c","Type":"ContainerDied","Data":"d6262d23e07e04bd7b2d55af127016e9cd5d355cc30c2f28b3012502078248de"} Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.784801 4857 scope.go:117] "RemoveContainer" containerID="96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.830242 4857 scope.go:117] "RemoveContainer" containerID="d74e2032540592afd088f7fa9601de559346100acee448f812a6080f4b577b85" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.845922 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9d749"] Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.858577 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9d749"] Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.889064 4857 scope.go:117] "RemoveContainer" containerID="43d6c1097a6cb03e1d0d353c1d396ff75759e65e2c1d8dd993636b7210c4ce0b" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.936197 4857 scope.go:117] "RemoveContainer" containerID="96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8" Dec 01 22:08:52 crc kubenswrapper[4857]: E1201 22:08:52.936936 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8\": container with ID starting with 96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8 not found: ID does not exist" containerID="96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.936990 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8"} err="failed to get container status \"96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8\": rpc error: code = NotFound desc = could not find container \"96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8\": container with ID starting with 96fa459bf5389668dcee2da66ccc67edc7399344c4e2abb746c7da8115c71df8 not found: ID does not exist" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.937072 4857 scope.go:117] "RemoveContainer" containerID="d74e2032540592afd088f7fa9601de559346100acee448f812a6080f4b577b85" Dec 01 22:08:52 crc kubenswrapper[4857]: E1201 22:08:52.937618 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d74e2032540592afd088f7fa9601de559346100acee448f812a6080f4b577b85\": container with ID starting with d74e2032540592afd088f7fa9601de559346100acee448f812a6080f4b577b85 not found: ID does not exist" containerID="d74e2032540592afd088f7fa9601de559346100acee448f812a6080f4b577b85" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.937658 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d74e2032540592afd088f7fa9601de559346100acee448f812a6080f4b577b85"} err="failed to get container status \"d74e2032540592afd088f7fa9601de559346100acee448f812a6080f4b577b85\": rpc error: code = NotFound desc = could not find container \"d74e2032540592afd088f7fa9601de559346100acee448f812a6080f4b577b85\": container with ID starting with d74e2032540592afd088f7fa9601de559346100acee448f812a6080f4b577b85 not found: ID does not exist" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.937682 4857 scope.go:117] "RemoveContainer" containerID="43d6c1097a6cb03e1d0d353c1d396ff75759e65e2c1d8dd993636b7210c4ce0b" Dec 01 22:08:52 crc kubenswrapper[4857]: E1201 22:08:52.938105 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43d6c1097a6cb03e1d0d353c1d396ff75759e65e2c1d8dd993636b7210c4ce0b\": container with ID starting with 43d6c1097a6cb03e1d0d353c1d396ff75759e65e2c1d8dd993636b7210c4ce0b not found: ID does not exist" containerID="43d6c1097a6cb03e1d0d353c1d396ff75759e65e2c1d8dd993636b7210c4ce0b" Dec 01 22:08:52 crc kubenswrapper[4857]: I1201 22:08:52.938137 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43d6c1097a6cb03e1d0d353c1d396ff75759e65e2c1d8dd993636b7210c4ce0b"} err="failed to get container status \"43d6c1097a6cb03e1d0d353c1d396ff75759e65e2c1d8dd993636b7210c4ce0b\": rpc error: code = NotFound desc = could not find container \"43d6c1097a6cb03e1d0d353c1d396ff75759e65e2c1d8dd993636b7210c4ce0b\": container with ID starting with 43d6c1097a6cb03e1d0d353c1d396ff75759e65e2c1d8dd993636b7210c4ce0b not found: ID does not exist" Dec 01 22:08:53 crc kubenswrapper[4857]: I1201 22:08:53.855958 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51b8e39d-30ef-4325-91ec-8b46219e093c" path="/var/lib/kubelet/pods/51b8e39d-30ef-4325-91ec-8b46219e093c/volumes" Dec 01 22:09:21 crc kubenswrapper[4857]: I1201 22:09:21.130246 4857 generic.go:334] "Generic (PLEG): container finished" podID="e3196a47-b0ed-4f5d-a639-496ff78d1acd" containerID="fdeb887b128fadfaaa9595b9e7060ea10bb5e0b77cf52a5e87510c9071bcb3f9" exitCode=0 Dec 01 22:09:21 crc kubenswrapper[4857]: I1201 22:09:21.130372 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" event={"ID":"e3196a47-b0ed-4f5d-a639-496ff78d1acd","Type":"ContainerDied","Data":"fdeb887b128fadfaaa9595b9e7060ea10bb5e0b77cf52a5e87510c9071bcb3f9"} Dec 01 22:09:22 crc kubenswrapper[4857]: I1201 22:09:22.701782 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:09:22 crc kubenswrapper[4857]: I1201 22:09:22.829863 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csgr8\" (UniqueName: \"kubernetes.io/projected/e3196a47-b0ed-4f5d-a639-496ff78d1acd-kube-api-access-csgr8\") pod \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\" (UID: \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\") " Dec 01 22:09:22 crc kubenswrapper[4857]: I1201 22:09:22.830228 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3196a47-b0ed-4f5d-a639-496ff78d1acd-inventory\") pod \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\" (UID: \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\") " Dec 01 22:09:22 crc kubenswrapper[4857]: I1201 22:09:22.831547 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3196a47-b0ed-4f5d-a639-496ff78d1acd-ssh-key\") pod \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\" (UID: \"e3196a47-b0ed-4f5d-a639-496ff78d1acd\") " Dec 01 22:09:22 crc kubenswrapper[4857]: I1201 22:09:22.836808 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3196a47-b0ed-4f5d-a639-496ff78d1acd-kube-api-access-csgr8" (OuterVolumeSpecName: "kube-api-access-csgr8") pod "e3196a47-b0ed-4f5d-a639-496ff78d1acd" (UID: "e3196a47-b0ed-4f5d-a639-496ff78d1acd"). InnerVolumeSpecName "kube-api-access-csgr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:09:22 crc kubenswrapper[4857]: I1201 22:09:22.859898 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3196a47-b0ed-4f5d-a639-496ff78d1acd-inventory" (OuterVolumeSpecName: "inventory") pod "e3196a47-b0ed-4f5d-a639-496ff78d1acd" (UID: "e3196a47-b0ed-4f5d-a639-496ff78d1acd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:09:22 crc kubenswrapper[4857]: I1201 22:09:22.888527 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3196a47-b0ed-4f5d-a639-496ff78d1acd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e3196a47-b0ed-4f5d-a639-496ff78d1acd" (UID: "e3196a47-b0ed-4f5d-a639-496ff78d1acd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:09:22 crc kubenswrapper[4857]: I1201 22:09:22.935193 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csgr8\" (UniqueName: \"kubernetes.io/projected/e3196a47-b0ed-4f5d-a639-496ff78d1acd-kube-api-access-csgr8\") on node \"crc\" DevicePath \"\"" Dec 01 22:09:22 crc kubenswrapper[4857]: I1201 22:09:22.935241 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3196a47-b0ed-4f5d-a639-496ff78d1acd-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:09:22 crc kubenswrapper[4857]: I1201 22:09:22.935261 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3196a47-b0ed-4f5d-a639-496ff78d1acd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.160097 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" event={"ID":"e3196a47-b0ed-4f5d-a639-496ff78d1acd","Type":"ContainerDied","Data":"2a977fe7e8ec3cc637226845aae8e9a59ac39aebc0671ade8d45d862df5bfe38"} Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.160157 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a977fe7e8ec3cc637226845aae8e9a59ac39aebc0671ade8d45d862df5bfe38" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.160213 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7mccz" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.275445 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8zsck"] Dec 01 22:09:23 crc kubenswrapper[4857]: E1201 22:09:23.275890 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3196a47-b0ed-4f5d-a639-496ff78d1acd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.275910 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3196a47-b0ed-4f5d-a639-496ff78d1acd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 22:09:23 crc kubenswrapper[4857]: E1201 22:09:23.275943 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b8e39d-30ef-4325-91ec-8b46219e093c" containerName="extract-utilities" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.275954 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b8e39d-30ef-4325-91ec-8b46219e093c" containerName="extract-utilities" Dec 01 22:09:23 crc kubenswrapper[4857]: E1201 22:09:23.275972 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b8e39d-30ef-4325-91ec-8b46219e093c" containerName="extract-content" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.275980 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b8e39d-30ef-4325-91ec-8b46219e093c" containerName="extract-content" Dec 01 22:09:23 crc kubenswrapper[4857]: E1201 22:09:23.276001 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b8e39d-30ef-4325-91ec-8b46219e093c" containerName="registry-server" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.276009 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b8e39d-30ef-4325-91ec-8b46219e093c" containerName="registry-server" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.276275 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3196a47-b0ed-4f5d-a639-496ff78d1acd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.276294 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b8e39d-30ef-4325-91ec-8b46219e093c" containerName="registry-server" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.277120 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.280985 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.281027 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.281129 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.282258 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.292180 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8zsck"] Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.446072 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7sp2\" (UniqueName: \"kubernetes.io/projected/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-kube-api-access-s7sp2\") pod \"ssh-known-hosts-edpm-deployment-8zsck\" (UID: \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.446395 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8zsck\" (UID: \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.446493 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8zsck\" (UID: \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.548534 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7sp2\" (UniqueName: \"kubernetes.io/projected/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-kube-api-access-s7sp2\") pod \"ssh-known-hosts-edpm-deployment-8zsck\" (UID: \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.548713 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8zsck\" (UID: \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.548769 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8zsck\" (UID: \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.555960 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8zsck\" (UID: \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.556458 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8zsck\" (UID: \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.571298 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7sp2\" (UniqueName: \"kubernetes.io/projected/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-kube-api-access-s7sp2\") pod \"ssh-known-hosts-edpm-deployment-8zsck\" (UID: \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:23 crc kubenswrapper[4857]: I1201 22:09:23.602129 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:24 crc kubenswrapper[4857]: I1201 22:09:24.243362 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8zsck"] Dec 01 22:09:24 crc kubenswrapper[4857]: I1201 22:09:24.252791 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 22:09:25 crc kubenswrapper[4857]: I1201 22:09:25.190941 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" event={"ID":"e7a9f9f9-0f72-46cc-9f30-461fe1a33903","Type":"ContainerStarted","Data":"2ae05e3d8f48f2e86d696b1a7ea8ab4f8d316e758a345932ec59c93e0e9e17f6"} Dec 01 22:09:25 crc kubenswrapper[4857]: I1201 22:09:25.191355 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" event={"ID":"e7a9f9f9-0f72-46cc-9f30-461fe1a33903","Type":"ContainerStarted","Data":"ea371463ef25e58351abc029979a02fa65f54c2c5625f66272c6af40f756a9f2"} Dec 01 22:09:25 crc kubenswrapper[4857]: I1201 22:09:25.226608 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" podStartSLOduration=1.732324577 podStartE2EDuration="2.226578099s" podCreationTimestamp="2025-12-01 22:09:23 +0000 UTC" firstStartedPulling="2025-12-01 22:09:24.252601234 +0000 UTC m=+2082.742663551" lastFinishedPulling="2025-12-01 22:09:24.746854756 +0000 UTC m=+2083.236917073" observedRunningTime="2025-12-01 22:09:25.212334557 +0000 UTC m=+2083.702396914" watchObservedRunningTime="2025-12-01 22:09:25.226578099 +0000 UTC m=+2083.716640446" Dec 01 22:09:33 crc kubenswrapper[4857]: I1201 22:09:33.291847 4857 generic.go:334] "Generic (PLEG): container finished" podID="e7a9f9f9-0f72-46cc-9f30-461fe1a33903" containerID="2ae05e3d8f48f2e86d696b1a7ea8ab4f8d316e758a345932ec59c93e0e9e17f6" exitCode=0 Dec 01 22:09:33 crc kubenswrapper[4857]: I1201 22:09:33.292269 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" event={"ID":"e7a9f9f9-0f72-46cc-9f30-461fe1a33903","Type":"ContainerDied","Data":"2ae05e3d8f48f2e86d696b1a7ea8ab4f8d316e758a345932ec59c93e0e9e17f6"} Dec 01 22:09:34 crc kubenswrapper[4857]: I1201 22:09:34.746818 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:34 crc kubenswrapper[4857]: I1201 22:09:34.832943 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-inventory-0\") pod \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\" (UID: \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\") " Dec 01 22:09:34 crc kubenswrapper[4857]: I1201 22:09:34.833085 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-ssh-key-openstack-edpm-ipam\") pod \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\" (UID: \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\") " Dec 01 22:09:34 crc kubenswrapper[4857]: I1201 22:09:34.833152 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7sp2\" (UniqueName: \"kubernetes.io/projected/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-kube-api-access-s7sp2\") pod \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\" (UID: \"e7a9f9f9-0f72-46cc-9f30-461fe1a33903\") " Dec 01 22:09:34 crc kubenswrapper[4857]: I1201 22:09:34.844428 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-kube-api-access-s7sp2" (OuterVolumeSpecName: "kube-api-access-s7sp2") pod "e7a9f9f9-0f72-46cc-9f30-461fe1a33903" (UID: "e7a9f9f9-0f72-46cc-9f30-461fe1a33903"). InnerVolumeSpecName "kube-api-access-s7sp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:09:34 crc kubenswrapper[4857]: I1201 22:09:34.860546 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e7a9f9f9-0f72-46cc-9f30-461fe1a33903" (UID: "e7a9f9f9-0f72-46cc-9f30-461fe1a33903"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:09:34 crc kubenswrapper[4857]: I1201 22:09:34.885183 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "e7a9f9f9-0f72-46cc-9f30-461fe1a33903" (UID: "e7a9f9f9-0f72-46cc-9f30-461fe1a33903"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:09:34 crc kubenswrapper[4857]: I1201 22:09:34.936404 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7sp2\" (UniqueName: \"kubernetes.io/projected/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-kube-api-access-s7sp2\") on node \"crc\" DevicePath \"\"" Dec 01 22:09:34 crc kubenswrapper[4857]: I1201 22:09:34.936581 4857 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:09:34 crc kubenswrapper[4857]: I1201 22:09:34.936718 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e7a9f9f9-0f72-46cc-9f30-461fe1a33903-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.316133 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" event={"ID":"e7a9f9f9-0f72-46cc-9f30-461fe1a33903","Type":"ContainerDied","Data":"ea371463ef25e58351abc029979a02fa65f54c2c5625f66272c6af40f756a9f2"} Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.316446 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea371463ef25e58351abc029979a02fa65f54c2c5625f66272c6af40f756a9f2" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.316267 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8zsck" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.420585 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h"] Dec 01 22:09:35 crc kubenswrapper[4857]: E1201 22:09:35.421197 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a9f9f9-0f72-46cc-9f30-461fe1a33903" containerName="ssh-known-hosts-edpm-deployment" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.421221 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a9f9f9-0f72-46cc-9f30-461fe1a33903" containerName="ssh-known-hosts-edpm-deployment" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.421536 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7a9f9f9-0f72-46cc-9f30-461fe1a33903" containerName="ssh-known-hosts-edpm-deployment" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.422579 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.430478 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h"] Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.477130 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.477314 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.477437 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.477796 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.479455 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnkqc\" (UniqueName: \"kubernetes.io/projected/a1548cec-634a-469e-b3bf-1813dfe676c1-kube-api-access-qnkqc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cwc7h\" (UID: \"a1548cec-634a-469e-b3bf-1813dfe676c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.479570 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1548cec-634a-469e-b3bf-1813dfe676c1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cwc7h\" (UID: \"a1548cec-634a-469e-b3bf-1813dfe676c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.479660 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1548cec-634a-469e-b3bf-1813dfe676c1-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cwc7h\" (UID: \"a1548cec-634a-469e-b3bf-1813dfe676c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.581850 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1548cec-634a-469e-b3bf-1813dfe676c1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cwc7h\" (UID: \"a1548cec-634a-469e-b3bf-1813dfe676c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.581990 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1548cec-634a-469e-b3bf-1813dfe676c1-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cwc7h\" (UID: \"a1548cec-634a-469e-b3bf-1813dfe676c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.582067 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnkqc\" (UniqueName: \"kubernetes.io/projected/a1548cec-634a-469e-b3bf-1813dfe676c1-kube-api-access-qnkqc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cwc7h\" (UID: \"a1548cec-634a-469e-b3bf-1813dfe676c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.586603 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1548cec-634a-469e-b3bf-1813dfe676c1-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cwc7h\" (UID: \"a1548cec-634a-469e-b3bf-1813dfe676c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.587680 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1548cec-634a-469e-b3bf-1813dfe676c1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cwc7h\" (UID: \"a1548cec-634a-469e-b3bf-1813dfe676c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.601417 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnkqc\" (UniqueName: \"kubernetes.io/projected/a1548cec-634a-469e-b3bf-1813dfe676c1-kube-api-access-qnkqc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cwc7h\" (UID: \"a1548cec-634a-469e-b3bf-1813dfe676c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:35 crc kubenswrapper[4857]: I1201 22:09:35.797034 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:36 crc kubenswrapper[4857]: I1201 22:09:36.355264 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h"] Dec 01 22:09:37 crc kubenswrapper[4857]: I1201 22:09:37.334872 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" event={"ID":"a1548cec-634a-469e-b3bf-1813dfe676c1","Type":"ContainerStarted","Data":"839836670aa332b0c2d09433146cc1d361e342ff112748c48b7813791befe8f6"} Dec 01 22:09:37 crc kubenswrapper[4857]: I1201 22:09:37.334929 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" event={"ID":"a1548cec-634a-469e-b3bf-1813dfe676c1","Type":"ContainerStarted","Data":"9bb4ad160ff14c6d6426ac877662d7753886dd97250c37409e3e80ee4c249218"} Dec 01 22:09:37 crc kubenswrapper[4857]: I1201 22:09:37.361501 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" podStartSLOduration=1.798667435 podStartE2EDuration="2.361479561s" podCreationTimestamp="2025-12-01 22:09:35 +0000 UTC" firstStartedPulling="2025-12-01 22:09:36.362587951 +0000 UTC m=+2094.852650268" lastFinishedPulling="2025-12-01 22:09:36.925400067 +0000 UTC m=+2095.415462394" observedRunningTime="2025-12-01 22:09:37.349912715 +0000 UTC m=+2095.839975052" watchObservedRunningTime="2025-12-01 22:09:37.361479561 +0000 UTC m=+2095.851541888" Dec 01 22:09:46 crc kubenswrapper[4857]: I1201 22:09:46.445552 4857 generic.go:334] "Generic (PLEG): container finished" podID="a1548cec-634a-469e-b3bf-1813dfe676c1" containerID="839836670aa332b0c2d09433146cc1d361e342ff112748c48b7813791befe8f6" exitCode=0 Dec 01 22:09:46 crc kubenswrapper[4857]: I1201 22:09:46.446260 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" event={"ID":"a1548cec-634a-469e-b3bf-1813dfe676c1","Type":"ContainerDied","Data":"839836670aa332b0c2d09433146cc1d361e342ff112748c48b7813791befe8f6"} Dec 01 22:09:47 crc kubenswrapper[4857]: I1201 22:09:47.845536 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.041460 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnkqc\" (UniqueName: \"kubernetes.io/projected/a1548cec-634a-469e-b3bf-1813dfe676c1-kube-api-access-qnkqc\") pod \"a1548cec-634a-469e-b3bf-1813dfe676c1\" (UID: \"a1548cec-634a-469e-b3bf-1813dfe676c1\") " Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.041656 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1548cec-634a-469e-b3bf-1813dfe676c1-inventory\") pod \"a1548cec-634a-469e-b3bf-1813dfe676c1\" (UID: \"a1548cec-634a-469e-b3bf-1813dfe676c1\") " Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.041736 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1548cec-634a-469e-b3bf-1813dfe676c1-ssh-key\") pod \"a1548cec-634a-469e-b3bf-1813dfe676c1\" (UID: \"a1548cec-634a-469e-b3bf-1813dfe676c1\") " Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.048443 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1548cec-634a-469e-b3bf-1813dfe676c1-kube-api-access-qnkqc" (OuterVolumeSpecName: "kube-api-access-qnkqc") pod "a1548cec-634a-469e-b3bf-1813dfe676c1" (UID: "a1548cec-634a-469e-b3bf-1813dfe676c1"). InnerVolumeSpecName "kube-api-access-qnkqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.069517 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1548cec-634a-469e-b3bf-1813dfe676c1-inventory" (OuterVolumeSpecName: "inventory") pod "a1548cec-634a-469e-b3bf-1813dfe676c1" (UID: "a1548cec-634a-469e-b3bf-1813dfe676c1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.074122 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1548cec-634a-469e-b3bf-1813dfe676c1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a1548cec-634a-469e-b3bf-1813dfe676c1" (UID: "a1548cec-634a-469e-b3bf-1813dfe676c1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.143911 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1548cec-634a-469e-b3bf-1813dfe676c1-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.144175 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1548cec-634a-469e-b3bf-1813dfe676c1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.144259 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnkqc\" (UniqueName: \"kubernetes.io/projected/a1548cec-634a-469e-b3bf-1813dfe676c1-kube-api-access-qnkqc\") on node \"crc\" DevicePath \"\"" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.476324 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" event={"ID":"a1548cec-634a-469e-b3bf-1813dfe676c1","Type":"ContainerDied","Data":"9bb4ad160ff14c6d6426ac877662d7753886dd97250c37409e3e80ee4c249218"} Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.476808 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bb4ad160ff14c6d6426ac877662d7753886dd97250c37409e3e80ee4c249218" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.476450 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cwc7h" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.629712 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4"] Dec 01 22:09:48 crc kubenswrapper[4857]: E1201 22:09:48.630197 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1548cec-634a-469e-b3bf-1813dfe676c1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.630220 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1548cec-634a-469e-b3bf-1813dfe676c1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.630460 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1548cec-634a-469e-b3bf-1813dfe676c1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.631156 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.633351 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.633904 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.634007 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.634354 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.646519 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4"] Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.763819 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9nll\" (UniqueName: \"kubernetes.io/projected/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-kube-api-access-p9nll\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4\" (UID: \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.763980 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4\" (UID: \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.764028 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4\" (UID: \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.865844 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4\" (UID: \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.865909 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4\" (UID: \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.866029 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9nll\" (UniqueName: \"kubernetes.io/projected/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-kube-api-access-p9nll\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4\" (UID: \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.872348 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4\" (UID: \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.873228 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4\" (UID: \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.894776 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9nll\" (UniqueName: \"kubernetes.io/projected/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-kube-api-access-p9nll\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4\" (UID: \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:09:48 crc kubenswrapper[4857]: I1201 22:09:48.954275 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:09:49 crc kubenswrapper[4857]: I1201 22:09:49.330248 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4"] Dec 01 22:09:49 crc kubenswrapper[4857]: I1201 22:09:49.487339 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" event={"ID":"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce","Type":"ContainerStarted","Data":"d8fe23b2b0512c898c26bdbd3f167ce00848296e678066065594b92e09218b13"} Dec 01 22:09:51 crc kubenswrapper[4857]: I1201 22:09:51.515703 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" event={"ID":"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce","Type":"ContainerStarted","Data":"4e2817f97d9699a74a06a99ab4f0c7887dc4ee2a59e7fca78611bdab831df7f6"} Dec 01 22:09:51 crc kubenswrapper[4857]: I1201 22:09:51.552818 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" podStartSLOduration=2.05442035 podStartE2EDuration="3.552790903s" podCreationTimestamp="2025-12-01 22:09:48 +0000 UTC" firstStartedPulling="2025-12-01 22:09:49.340551442 +0000 UTC m=+2107.830613759" lastFinishedPulling="2025-12-01 22:09:50.838921965 +0000 UTC m=+2109.328984312" observedRunningTime="2025-12-01 22:09:51.534824109 +0000 UTC m=+2110.024886476" watchObservedRunningTime="2025-12-01 22:09:51.552790903 +0000 UTC m=+2110.042853260" Dec 01 22:10:01 crc kubenswrapper[4857]: I1201 22:10:01.618581 4857 generic.go:334] "Generic (PLEG): container finished" podID="ad6a5aa6-3dc8-4387-9834-7681a0bf77ce" containerID="4e2817f97d9699a74a06a99ab4f0c7887dc4ee2a59e7fca78611bdab831df7f6" exitCode=0 Dec 01 22:10:01 crc kubenswrapper[4857]: I1201 22:10:01.618682 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" event={"ID":"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce","Type":"ContainerDied","Data":"4e2817f97d9699a74a06a99ab4f0c7887dc4ee2a59e7fca78611bdab831df7f6"} Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.043038 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.124195 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-ssh-key\") pod \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\" (UID: \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\") " Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.124291 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9nll\" (UniqueName: \"kubernetes.io/projected/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-kube-api-access-p9nll\") pod \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\" (UID: \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\") " Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.124374 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-inventory\") pod \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\" (UID: \"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce\") " Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.131801 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-kube-api-access-p9nll" (OuterVolumeSpecName: "kube-api-access-p9nll") pod "ad6a5aa6-3dc8-4387-9834-7681a0bf77ce" (UID: "ad6a5aa6-3dc8-4387-9834-7681a0bf77ce"). InnerVolumeSpecName "kube-api-access-p9nll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.166186 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ad6a5aa6-3dc8-4387-9834-7681a0bf77ce" (UID: "ad6a5aa6-3dc8-4387-9834-7681a0bf77ce"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.195204 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-inventory" (OuterVolumeSpecName: "inventory") pod "ad6a5aa6-3dc8-4387-9834-7681a0bf77ce" (UID: "ad6a5aa6-3dc8-4387-9834-7681a0bf77ce"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.226456 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.226488 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.226498 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9nll\" (UniqueName: \"kubernetes.io/projected/ad6a5aa6-3dc8-4387-9834-7681a0bf77ce-kube-api-access-p9nll\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.636546 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" event={"ID":"ad6a5aa6-3dc8-4387-9834-7681a0bf77ce","Type":"ContainerDied","Data":"d8fe23b2b0512c898c26bdbd3f167ce00848296e678066065594b92e09218b13"} Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.636598 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8fe23b2b0512c898c26bdbd3f167ce00848296e678066065594b92e09218b13" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.636694 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.734015 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp"] Dec 01 22:10:03 crc kubenswrapper[4857]: E1201 22:10:03.734442 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad6a5aa6-3dc8-4387-9834-7681a0bf77ce" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.734464 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad6a5aa6-3dc8-4387-9834-7681a0bf77ce" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.734666 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad6a5aa6-3dc8-4387-9834-7681a0bf77ce" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.735391 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.737751 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.738640 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.743614 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.743632 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.743776 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.743827 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.743891 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.744002 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.748768 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp"] Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.837298 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.837457 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.837524 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h57jc\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-kube-api-access-h57jc\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.837582 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.837624 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.837828 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.837931 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.837999 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.838309 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.838378 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.838538 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.838627 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.838697 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.838756 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.940252 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.940332 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.940471 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.940516 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.940569 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.940606 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.940641 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.940762 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.940800 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h57jc\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-kube-api-access-h57jc\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.940834 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.940863 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.941037 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.941124 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.941174 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.948847 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.949404 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.949998 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.956534 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.957418 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.962665 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.963743 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.964186 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.964464 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.965176 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.965821 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.967103 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.973079 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:03 crc kubenswrapper[4857]: I1201 22:10:03.980582 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h57jc\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-kube-api-access-h57jc\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:04 crc kubenswrapper[4857]: I1201 22:10:04.090931 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:04 crc kubenswrapper[4857]: I1201 22:10:04.700302 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp"] Dec 01 22:10:04 crc kubenswrapper[4857]: W1201 22:10:04.706773 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf71f02cd_2132_437a_b18c_e3513eb0710c.slice/crio-9085ecd24ff1a1cf3562cfd105dabacb4656afb38ecdf12f352e5ad5ce37447a WatchSource:0}: Error finding container 9085ecd24ff1a1cf3562cfd105dabacb4656afb38ecdf12f352e5ad5ce37447a: Status 404 returned error can't find the container with id 9085ecd24ff1a1cf3562cfd105dabacb4656afb38ecdf12f352e5ad5ce37447a Dec 01 22:10:05 crc kubenswrapper[4857]: I1201 22:10:05.660366 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" event={"ID":"f71f02cd-2132-437a-b18c-e3513eb0710c","Type":"ContainerStarted","Data":"9085ecd24ff1a1cf3562cfd105dabacb4656afb38ecdf12f352e5ad5ce37447a"} Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.376256 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kq2bz"] Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.378434 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.412001 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kq2bz"] Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.499608 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/679de64c-96c8-4ed0-b236-1c8f93fba695-catalog-content\") pod \"redhat-marketplace-kq2bz\" (UID: \"679de64c-96c8-4ed0-b236-1c8f93fba695\") " pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.499780 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/679de64c-96c8-4ed0-b236-1c8f93fba695-utilities\") pod \"redhat-marketplace-kq2bz\" (UID: \"679de64c-96c8-4ed0-b236-1c8f93fba695\") " pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.499993 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86qd8\" (UniqueName: \"kubernetes.io/projected/679de64c-96c8-4ed0-b236-1c8f93fba695-kube-api-access-86qd8\") pod \"redhat-marketplace-kq2bz\" (UID: \"679de64c-96c8-4ed0-b236-1c8f93fba695\") " pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.601458 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/679de64c-96c8-4ed0-b236-1c8f93fba695-catalog-content\") pod \"redhat-marketplace-kq2bz\" (UID: \"679de64c-96c8-4ed0-b236-1c8f93fba695\") " pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.601600 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/679de64c-96c8-4ed0-b236-1c8f93fba695-utilities\") pod \"redhat-marketplace-kq2bz\" (UID: \"679de64c-96c8-4ed0-b236-1c8f93fba695\") " pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.602213 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/679de64c-96c8-4ed0-b236-1c8f93fba695-catalog-content\") pod \"redhat-marketplace-kq2bz\" (UID: \"679de64c-96c8-4ed0-b236-1c8f93fba695\") " pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.602287 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86qd8\" (UniqueName: \"kubernetes.io/projected/679de64c-96c8-4ed0-b236-1c8f93fba695-kube-api-access-86qd8\") pod \"redhat-marketplace-kq2bz\" (UID: \"679de64c-96c8-4ed0-b236-1c8f93fba695\") " pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.602262 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/679de64c-96c8-4ed0-b236-1c8f93fba695-utilities\") pod \"redhat-marketplace-kq2bz\" (UID: \"679de64c-96c8-4ed0-b236-1c8f93fba695\") " pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.619816 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86qd8\" (UniqueName: \"kubernetes.io/projected/679de64c-96c8-4ed0-b236-1c8f93fba695-kube-api-access-86qd8\") pod \"redhat-marketplace-kq2bz\" (UID: \"679de64c-96c8-4ed0-b236-1c8f93fba695\") " pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.673251 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" event={"ID":"f71f02cd-2132-437a-b18c-e3513eb0710c","Type":"ContainerStarted","Data":"a793d07ceacb653dd2b3abd7c379f6cf2ef76f073618c194096615073d43b44b"} Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.699937 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:06 crc kubenswrapper[4857]: I1201 22:10:06.702316 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" podStartSLOduration=2.690592 podStartE2EDuration="3.702294017s" podCreationTimestamp="2025-12-01 22:10:03 +0000 UTC" firstStartedPulling="2025-12-01 22:10:04.710471925 +0000 UTC m=+2123.200534252" lastFinishedPulling="2025-12-01 22:10:05.722173922 +0000 UTC m=+2124.212236269" observedRunningTime="2025-12-01 22:10:06.694541336 +0000 UTC m=+2125.184603653" watchObservedRunningTime="2025-12-01 22:10:06.702294017 +0000 UTC m=+2125.192356334" Dec 01 22:10:07 crc kubenswrapper[4857]: I1201 22:10:07.215175 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kq2bz"] Dec 01 22:10:07 crc kubenswrapper[4857]: I1201 22:10:07.681704 4857 generic.go:334] "Generic (PLEG): container finished" podID="679de64c-96c8-4ed0-b236-1c8f93fba695" containerID="5fd8468ff27e4fdf18cd311987e9e88ec63bd9a57a54070143a3bd157033b646" exitCode=0 Dec 01 22:10:07 crc kubenswrapper[4857]: I1201 22:10:07.681877 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kq2bz" event={"ID":"679de64c-96c8-4ed0-b236-1c8f93fba695","Type":"ContainerDied","Data":"5fd8468ff27e4fdf18cd311987e9e88ec63bd9a57a54070143a3bd157033b646"} Dec 01 22:10:07 crc kubenswrapper[4857]: I1201 22:10:07.682292 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kq2bz" event={"ID":"679de64c-96c8-4ed0-b236-1c8f93fba695","Type":"ContainerStarted","Data":"62f51acb03c8c9bfed942a4322dfa9a5c57081859b1ca238f618c2bdd1ed5c06"} Dec 01 22:10:08 crc kubenswrapper[4857]: I1201 22:10:08.692147 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kq2bz" event={"ID":"679de64c-96c8-4ed0-b236-1c8f93fba695","Type":"ContainerStarted","Data":"99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9"} Dec 01 22:10:08 crc kubenswrapper[4857]: E1201 22:10:08.893458 4857 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod679de64c_96c8_4ed0_b236_1c8f93fba695.slice/crio-99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9.scope\": RecentStats: unable to find data in memory cache]" Dec 01 22:10:09 crc kubenswrapper[4857]: I1201 22:10:09.702257 4857 generic.go:334] "Generic (PLEG): container finished" podID="679de64c-96c8-4ed0-b236-1c8f93fba695" containerID="99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9" exitCode=0 Dec 01 22:10:09 crc kubenswrapper[4857]: I1201 22:10:09.702345 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kq2bz" event={"ID":"679de64c-96c8-4ed0-b236-1c8f93fba695","Type":"ContainerDied","Data":"99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9"} Dec 01 22:10:10 crc kubenswrapper[4857]: I1201 22:10:10.714098 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kq2bz" event={"ID":"679de64c-96c8-4ed0-b236-1c8f93fba695","Type":"ContainerStarted","Data":"65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac"} Dec 01 22:10:10 crc kubenswrapper[4857]: I1201 22:10:10.734495 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kq2bz" podStartSLOduration=2.289573865 podStartE2EDuration="4.734465122s" podCreationTimestamp="2025-12-01 22:10:06 +0000 UTC" firstStartedPulling="2025-12-01 22:10:07.683800608 +0000 UTC m=+2126.173862925" lastFinishedPulling="2025-12-01 22:10:10.128691825 +0000 UTC m=+2128.618754182" observedRunningTime="2025-12-01 22:10:10.732095203 +0000 UTC m=+2129.222157530" watchObservedRunningTime="2025-12-01 22:10:10.734465122 +0000 UTC m=+2129.224527449" Dec 01 22:10:16 crc kubenswrapper[4857]: I1201 22:10:16.701182 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:16 crc kubenswrapper[4857]: I1201 22:10:16.701807 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:16 crc kubenswrapper[4857]: I1201 22:10:16.763283 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:16 crc kubenswrapper[4857]: I1201 22:10:16.850650 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:17 crc kubenswrapper[4857]: I1201 22:10:17.019644 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kq2bz"] Dec 01 22:10:18 crc kubenswrapper[4857]: I1201 22:10:18.794940 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kq2bz" podUID="679de64c-96c8-4ed0-b236-1c8f93fba695" containerName="registry-server" containerID="cri-o://65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac" gracePeriod=2 Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.329495 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.390550 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/679de64c-96c8-4ed0-b236-1c8f93fba695-catalog-content\") pod \"679de64c-96c8-4ed0-b236-1c8f93fba695\" (UID: \"679de64c-96c8-4ed0-b236-1c8f93fba695\") " Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.390701 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86qd8\" (UniqueName: \"kubernetes.io/projected/679de64c-96c8-4ed0-b236-1c8f93fba695-kube-api-access-86qd8\") pod \"679de64c-96c8-4ed0-b236-1c8f93fba695\" (UID: \"679de64c-96c8-4ed0-b236-1c8f93fba695\") " Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.390762 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/679de64c-96c8-4ed0-b236-1c8f93fba695-utilities\") pod \"679de64c-96c8-4ed0-b236-1c8f93fba695\" (UID: \"679de64c-96c8-4ed0-b236-1c8f93fba695\") " Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.392104 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/679de64c-96c8-4ed0-b236-1c8f93fba695-utilities" (OuterVolumeSpecName: "utilities") pod "679de64c-96c8-4ed0-b236-1c8f93fba695" (UID: "679de64c-96c8-4ed0-b236-1c8f93fba695"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.489008 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/679de64c-96c8-4ed0-b236-1c8f93fba695-kube-api-access-86qd8" (OuterVolumeSpecName: "kube-api-access-86qd8") pod "679de64c-96c8-4ed0-b236-1c8f93fba695" (UID: "679de64c-96c8-4ed0-b236-1c8f93fba695"). InnerVolumeSpecName "kube-api-access-86qd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.493521 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86qd8\" (UniqueName: \"kubernetes.io/projected/679de64c-96c8-4ed0-b236-1c8f93fba695-kube-api-access-86qd8\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.493548 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/679de64c-96c8-4ed0-b236-1c8f93fba695-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.502321 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/679de64c-96c8-4ed0-b236-1c8f93fba695-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "679de64c-96c8-4ed0-b236-1c8f93fba695" (UID: "679de64c-96c8-4ed0-b236-1c8f93fba695"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.595656 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/679de64c-96c8-4ed0-b236-1c8f93fba695-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.805507 4857 generic.go:334] "Generic (PLEG): container finished" podID="679de64c-96c8-4ed0-b236-1c8f93fba695" containerID="65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac" exitCode=0 Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.805562 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kq2bz" event={"ID":"679de64c-96c8-4ed0-b236-1c8f93fba695","Type":"ContainerDied","Data":"65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac"} Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.805618 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kq2bz" event={"ID":"679de64c-96c8-4ed0-b236-1c8f93fba695","Type":"ContainerDied","Data":"62f51acb03c8c9bfed942a4322dfa9a5c57081859b1ca238f618c2bdd1ed5c06"} Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.805652 4857 scope.go:117] "RemoveContainer" containerID="65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.806591 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kq2bz" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.842612 4857 scope.go:117] "RemoveContainer" containerID="99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.877523 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kq2bz"] Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.890178 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kq2bz"] Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.890461 4857 scope.go:117] "RemoveContainer" containerID="5fd8468ff27e4fdf18cd311987e9e88ec63bd9a57a54070143a3bd157033b646" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.955232 4857 scope.go:117] "RemoveContainer" containerID="65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac" Dec 01 22:10:19 crc kubenswrapper[4857]: E1201 22:10:19.956137 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac\": container with ID starting with 65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac not found: ID does not exist" containerID="65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.956206 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac"} err="failed to get container status \"65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac\": rpc error: code = NotFound desc = could not find container \"65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac\": container with ID starting with 65931bedb36ac95fc04f93a14910f943ffe5d8bc82a4b645a287a5ea7c99a0ac not found: ID does not exist" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.956249 4857 scope.go:117] "RemoveContainer" containerID="99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9" Dec 01 22:10:19 crc kubenswrapper[4857]: E1201 22:10:19.956698 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9\": container with ID starting with 99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9 not found: ID does not exist" containerID="99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.956759 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9"} err="failed to get container status \"99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9\": rpc error: code = NotFound desc = could not find container \"99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9\": container with ID starting with 99a9ad437ee6dc47287092fde03ea3fd6eef8362b84a4131304ce98cd61183d9 not found: ID does not exist" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.956806 4857 scope.go:117] "RemoveContainer" containerID="5fd8468ff27e4fdf18cd311987e9e88ec63bd9a57a54070143a3bd157033b646" Dec 01 22:10:19 crc kubenswrapper[4857]: E1201 22:10:19.957302 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fd8468ff27e4fdf18cd311987e9e88ec63bd9a57a54070143a3bd157033b646\": container with ID starting with 5fd8468ff27e4fdf18cd311987e9e88ec63bd9a57a54070143a3bd157033b646 not found: ID does not exist" containerID="5fd8468ff27e4fdf18cd311987e9e88ec63bd9a57a54070143a3bd157033b646" Dec 01 22:10:19 crc kubenswrapper[4857]: I1201 22:10:19.957337 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fd8468ff27e4fdf18cd311987e9e88ec63bd9a57a54070143a3bd157033b646"} err="failed to get container status \"5fd8468ff27e4fdf18cd311987e9e88ec63bd9a57a54070143a3bd157033b646\": rpc error: code = NotFound desc = could not find container \"5fd8468ff27e4fdf18cd311987e9e88ec63bd9a57a54070143a3bd157033b646\": container with ID starting with 5fd8468ff27e4fdf18cd311987e9e88ec63bd9a57a54070143a3bd157033b646 not found: ID does not exist" Dec 01 22:10:21 crc kubenswrapper[4857]: I1201 22:10:21.853445 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="679de64c-96c8-4ed0-b236-1c8f93fba695" path="/var/lib/kubelet/pods/679de64c-96c8-4ed0-b236-1c8f93fba695/volumes" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.634377 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lfw2b"] Dec 01 22:10:26 crc kubenswrapper[4857]: E1201 22:10:26.635567 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="679de64c-96c8-4ed0-b236-1c8f93fba695" containerName="registry-server" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.635592 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="679de64c-96c8-4ed0-b236-1c8f93fba695" containerName="registry-server" Dec 01 22:10:26 crc kubenswrapper[4857]: E1201 22:10:26.635623 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="679de64c-96c8-4ed0-b236-1c8f93fba695" containerName="extract-utilities" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.635638 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="679de64c-96c8-4ed0-b236-1c8f93fba695" containerName="extract-utilities" Dec 01 22:10:26 crc kubenswrapper[4857]: E1201 22:10:26.635662 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="679de64c-96c8-4ed0-b236-1c8f93fba695" containerName="extract-content" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.635674 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="679de64c-96c8-4ed0-b236-1c8f93fba695" containerName="extract-content" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.636015 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="679de64c-96c8-4ed0-b236-1c8f93fba695" containerName="registry-server" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.638606 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.691465 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lfw2b"] Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.718224 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3106c12-2d34-416c-ae4f-d30b70b5bed3-catalog-content\") pod \"community-operators-lfw2b\" (UID: \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\") " pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.718324 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfbmh\" (UniqueName: \"kubernetes.io/projected/c3106c12-2d34-416c-ae4f-d30b70b5bed3-kube-api-access-tfbmh\") pod \"community-operators-lfw2b\" (UID: \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\") " pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.718389 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3106c12-2d34-416c-ae4f-d30b70b5bed3-utilities\") pod \"community-operators-lfw2b\" (UID: \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\") " pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.820225 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfbmh\" (UniqueName: \"kubernetes.io/projected/c3106c12-2d34-416c-ae4f-d30b70b5bed3-kube-api-access-tfbmh\") pod \"community-operators-lfw2b\" (UID: \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\") " pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.820614 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3106c12-2d34-416c-ae4f-d30b70b5bed3-utilities\") pod \"community-operators-lfw2b\" (UID: \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\") " pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.820743 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3106c12-2d34-416c-ae4f-d30b70b5bed3-catalog-content\") pod \"community-operators-lfw2b\" (UID: \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\") " pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.821158 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3106c12-2d34-416c-ae4f-d30b70b5bed3-utilities\") pod \"community-operators-lfw2b\" (UID: \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\") " pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.821211 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3106c12-2d34-416c-ae4f-d30b70b5bed3-catalog-content\") pod \"community-operators-lfw2b\" (UID: \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\") " pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.853994 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfbmh\" (UniqueName: \"kubernetes.io/projected/c3106c12-2d34-416c-ae4f-d30b70b5bed3-kube-api-access-tfbmh\") pod \"community-operators-lfw2b\" (UID: \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\") " pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:26 crc kubenswrapper[4857]: I1201 22:10:26.973611 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:27 crc kubenswrapper[4857]: I1201 22:10:27.319822 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lfw2b"] Dec 01 22:10:27 crc kubenswrapper[4857]: I1201 22:10:27.806082 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:10:27 crc kubenswrapper[4857]: I1201 22:10:27.806511 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:10:27 crc kubenswrapper[4857]: I1201 22:10:27.882922 4857 generic.go:334] "Generic (PLEG): container finished" podID="c3106c12-2d34-416c-ae4f-d30b70b5bed3" containerID="33744ac903a3021c97152204f9ff88859f3ec9b228e318af4bdddb8ba2740ba4" exitCode=0 Dec 01 22:10:27 crc kubenswrapper[4857]: I1201 22:10:27.882965 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfw2b" event={"ID":"c3106c12-2d34-416c-ae4f-d30b70b5bed3","Type":"ContainerDied","Data":"33744ac903a3021c97152204f9ff88859f3ec9b228e318af4bdddb8ba2740ba4"} Dec 01 22:10:27 crc kubenswrapper[4857]: I1201 22:10:27.882992 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfw2b" event={"ID":"c3106c12-2d34-416c-ae4f-d30b70b5bed3","Type":"ContainerStarted","Data":"e349d37447a453e8f6bfa011784f1aa4bd1c24b32d26719ee7d91bfcc44cc0a5"} Dec 01 22:10:28 crc kubenswrapper[4857]: I1201 22:10:28.909663 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfw2b" event={"ID":"c3106c12-2d34-416c-ae4f-d30b70b5bed3","Type":"ContainerStarted","Data":"39def8848d490001967f0358c11385a3a292435997e7e5272dedfbb8bcda61cc"} Dec 01 22:10:29 crc kubenswrapper[4857]: I1201 22:10:29.923573 4857 generic.go:334] "Generic (PLEG): container finished" podID="c3106c12-2d34-416c-ae4f-d30b70b5bed3" containerID="39def8848d490001967f0358c11385a3a292435997e7e5272dedfbb8bcda61cc" exitCode=0 Dec 01 22:10:29 crc kubenswrapper[4857]: I1201 22:10:29.923661 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfw2b" event={"ID":"c3106c12-2d34-416c-ae4f-d30b70b5bed3","Type":"ContainerDied","Data":"39def8848d490001967f0358c11385a3a292435997e7e5272dedfbb8bcda61cc"} Dec 01 22:10:30 crc kubenswrapper[4857]: I1201 22:10:30.938611 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfw2b" event={"ID":"c3106c12-2d34-416c-ae4f-d30b70b5bed3","Type":"ContainerStarted","Data":"99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349"} Dec 01 22:10:30 crc kubenswrapper[4857]: I1201 22:10:30.973193 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lfw2b" podStartSLOduration=2.211245746 podStartE2EDuration="4.973173064s" podCreationTimestamp="2025-12-01 22:10:26 +0000 UTC" firstStartedPulling="2025-12-01 22:10:27.885898986 +0000 UTC m=+2146.375961343" lastFinishedPulling="2025-12-01 22:10:30.647826324 +0000 UTC m=+2149.137888661" observedRunningTime="2025-12-01 22:10:30.972529509 +0000 UTC m=+2149.462591886" watchObservedRunningTime="2025-12-01 22:10:30.973173064 +0000 UTC m=+2149.463235391" Dec 01 22:10:36 crc kubenswrapper[4857]: I1201 22:10:36.973827 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:36 crc kubenswrapper[4857]: I1201 22:10:36.974510 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:37 crc kubenswrapper[4857]: I1201 22:10:37.050250 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:37 crc kubenswrapper[4857]: I1201 22:10:37.138441 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:37 crc kubenswrapper[4857]: I1201 22:10:37.591115 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lfw2b"] Dec 01 22:10:39 crc kubenswrapper[4857]: I1201 22:10:39.016571 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lfw2b" podUID="c3106c12-2d34-416c-ae4f-d30b70b5bed3" containerName="registry-server" containerID="cri-o://99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349" gracePeriod=2 Dec 01 22:10:39 crc kubenswrapper[4857]: I1201 22:10:39.701420 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:39 crc kubenswrapper[4857]: I1201 22:10:39.907079 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3106c12-2d34-416c-ae4f-d30b70b5bed3-utilities\") pod \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\" (UID: \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\") " Dec 01 22:10:39 crc kubenswrapper[4857]: I1201 22:10:39.907828 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfbmh\" (UniqueName: \"kubernetes.io/projected/c3106c12-2d34-416c-ae4f-d30b70b5bed3-kube-api-access-tfbmh\") pod \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\" (UID: \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\") " Dec 01 22:10:39 crc kubenswrapper[4857]: I1201 22:10:39.908154 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3106c12-2d34-416c-ae4f-d30b70b5bed3-catalog-content\") pod \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\" (UID: \"c3106c12-2d34-416c-ae4f-d30b70b5bed3\") " Dec 01 22:10:39 crc kubenswrapper[4857]: I1201 22:10:39.908667 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3106c12-2d34-416c-ae4f-d30b70b5bed3-utilities" (OuterVolumeSpecName: "utilities") pod "c3106c12-2d34-416c-ae4f-d30b70b5bed3" (UID: "c3106c12-2d34-416c-ae4f-d30b70b5bed3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:10:39 crc kubenswrapper[4857]: I1201 22:10:39.909919 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3106c12-2d34-416c-ae4f-d30b70b5bed3-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:39 crc kubenswrapper[4857]: I1201 22:10:39.915539 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3106c12-2d34-416c-ae4f-d30b70b5bed3-kube-api-access-tfbmh" (OuterVolumeSpecName: "kube-api-access-tfbmh") pod "c3106c12-2d34-416c-ae4f-d30b70b5bed3" (UID: "c3106c12-2d34-416c-ae4f-d30b70b5bed3"). InnerVolumeSpecName "kube-api-access-tfbmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:10:39 crc kubenswrapper[4857]: I1201 22:10:39.987547 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3106c12-2d34-416c-ae4f-d30b70b5bed3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3106c12-2d34-416c-ae4f-d30b70b5bed3" (UID: "c3106c12-2d34-416c-ae4f-d30b70b5bed3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.012101 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3106c12-2d34-416c-ae4f-d30b70b5bed3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.012329 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfbmh\" (UniqueName: \"kubernetes.io/projected/c3106c12-2d34-416c-ae4f-d30b70b5bed3-kube-api-access-tfbmh\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.031520 4857 generic.go:334] "Generic (PLEG): container finished" podID="c3106c12-2d34-416c-ae4f-d30b70b5bed3" containerID="99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349" exitCode=0 Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.031585 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfw2b" event={"ID":"c3106c12-2d34-416c-ae4f-d30b70b5bed3","Type":"ContainerDied","Data":"99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349"} Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.031626 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfw2b" event={"ID":"c3106c12-2d34-416c-ae4f-d30b70b5bed3","Type":"ContainerDied","Data":"e349d37447a453e8f6bfa011784f1aa4bd1c24b32d26719ee7d91bfcc44cc0a5"} Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.031655 4857 scope.go:117] "RemoveContainer" containerID="99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349" Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.031851 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfw2b" Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.078668 4857 scope.go:117] "RemoveContainer" containerID="39def8848d490001967f0358c11385a3a292435997e7e5272dedfbb8bcda61cc" Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.081399 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lfw2b"] Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.091342 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lfw2b"] Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.111739 4857 scope.go:117] "RemoveContainer" containerID="33744ac903a3021c97152204f9ff88859f3ec9b228e318af4bdddb8ba2740ba4" Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.163896 4857 scope.go:117] "RemoveContainer" containerID="99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349" Dec 01 22:10:40 crc kubenswrapper[4857]: E1201 22:10:40.164895 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349\": container with ID starting with 99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349 not found: ID does not exist" containerID="99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349" Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.165029 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349"} err="failed to get container status \"99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349\": rpc error: code = NotFound desc = could not find container \"99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349\": container with ID starting with 99b1db4d73c00226eaea7ca1360c464b855395cf8574897f26f007716a231349 not found: ID does not exist" Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.165165 4857 scope.go:117] "RemoveContainer" containerID="39def8848d490001967f0358c11385a3a292435997e7e5272dedfbb8bcda61cc" Dec 01 22:10:40 crc kubenswrapper[4857]: E1201 22:10:40.165744 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39def8848d490001967f0358c11385a3a292435997e7e5272dedfbb8bcda61cc\": container with ID starting with 39def8848d490001967f0358c11385a3a292435997e7e5272dedfbb8bcda61cc not found: ID does not exist" containerID="39def8848d490001967f0358c11385a3a292435997e7e5272dedfbb8bcda61cc" Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.165792 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39def8848d490001967f0358c11385a3a292435997e7e5272dedfbb8bcda61cc"} err="failed to get container status \"39def8848d490001967f0358c11385a3a292435997e7e5272dedfbb8bcda61cc\": rpc error: code = NotFound desc = could not find container \"39def8848d490001967f0358c11385a3a292435997e7e5272dedfbb8bcda61cc\": container with ID starting with 39def8848d490001967f0358c11385a3a292435997e7e5272dedfbb8bcda61cc not found: ID does not exist" Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.165825 4857 scope.go:117] "RemoveContainer" containerID="33744ac903a3021c97152204f9ff88859f3ec9b228e318af4bdddb8ba2740ba4" Dec 01 22:10:40 crc kubenswrapper[4857]: E1201 22:10:40.166225 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33744ac903a3021c97152204f9ff88859f3ec9b228e318af4bdddb8ba2740ba4\": container with ID starting with 33744ac903a3021c97152204f9ff88859f3ec9b228e318af4bdddb8ba2740ba4 not found: ID does not exist" containerID="33744ac903a3021c97152204f9ff88859f3ec9b228e318af4bdddb8ba2740ba4" Dec 01 22:10:40 crc kubenswrapper[4857]: I1201 22:10:40.166262 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33744ac903a3021c97152204f9ff88859f3ec9b228e318af4bdddb8ba2740ba4"} err="failed to get container status \"33744ac903a3021c97152204f9ff88859f3ec9b228e318af4bdddb8ba2740ba4\": rpc error: code = NotFound desc = could not find container \"33744ac903a3021c97152204f9ff88859f3ec9b228e318af4bdddb8ba2740ba4\": container with ID starting with 33744ac903a3021c97152204f9ff88859f3ec9b228e318af4bdddb8ba2740ba4 not found: ID does not exist" Dec 01 22:10:41 crc kubenswrapper[4857]: I1201 22:10:41.851779 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3106c12-2d34-416c-ae4f-d30b70b5bed3" path="/var/lib/kubelet/pods/c3106c12-2d34-416c-ae4f-d30b70b5bed3/volumes" Dec 01 22:10:51 crc kubenswrapper[4857]: I1201 22:10:51.148798 4857 generic.go:334] "Generic (PLEG): container finished" podID="f71f02cd-2132-437a-b18c-e3513eb0710c" containerID="a793d07ceacb653dd2b3abd7c379f6cf2ef76f073618c194096615073d43b44b" exitCode=0 Dec 01 22:10:51 crc kubenswrapper[4857]: I1201 22:10:51.148913 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" event={"ID":"f71f02cd-2132-437a-b18c-e3513eb0710c","Type":"ContainerDied","Data":"a793d07ceacb653dd2b3abd7c379f6cf2ef76f073618c194096615073d43b44b"} Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.728937 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.841953 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842190 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-ovn-combined-ca-bundle\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842275 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842299 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-nova-combined-ca-bundle\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842329 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-telemetry-combined-ca-bundle\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842352 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-repo-setup-combined-ca-bundle\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842382 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-libvirt-combined-ca-bundle\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842445 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-inventory\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842477 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842516 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-ssh-key\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842540 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-bootstrap-combined-ca-bundle\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842626 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h57jc\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-kube-api-access-h57jc\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842674 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-neutron-metadata-combined-ca-bundle\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.842714 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"f71f02cd-2132-437a-b18c-e3513eb0710c\" (UID: \"f71f02cd-2132-437a-b18c-e3513eb0710c\") " Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.851092 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.851752 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.851791 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.852666 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.853310 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.853446 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.853523 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-kube-api-access-h57jc" (OuterVolumeSpecName: "kube-api-access-h57jc") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "kube-api-access-h57jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.854981 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.855498 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.855801 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.856957 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.857775 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.874250 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-inventory" (OuterVolumeSpecName: "inventory") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.882577 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f71f02cd-2132-437a-b18c-e3513eb0710c" (UID: "f71f02cd-2132-437a-b18c-e3513eb0710c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945206 4857 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945242 4857 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945253 4857 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945263 4857 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945273 4857 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945281 4857 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945291 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945299 4857 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945309 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945318 4857 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945328 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h57jc\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-kube-api-access-h57jc\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945337 4857 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71f02cd-2132-437a-b18c-e3513eb0710c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945346 4857 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:52 crc kubenswrapper[4857]: I1201 22:10:52.945358 4857 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f71f02cd-2132-437a-b18c-e3513eb0710c-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.177853 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" event={"ID":"f71f02cd-2132-437a-b18c-e3513eb0710c","Type":"ContainerDied","Data":"9085ecd24ff1a1cf3562cfd105dabacb4656afb38ecdf12f352e5ad5ce37447a"} Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.177915 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9085ecd24ff1a1cf3562cfd105dabacb4656afb38ecdf12f352e5ad5ce37447a" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.177951 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.410098 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk"] Dec 01 22:10:53 crc kubenswrapper[4857]: E1201 22:10:53.410673 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3106c12-2d34-416c-ae4f-d30b70b5bed3" containerName="extract-content" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.410686 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3106c12-2d34-416c-ae4f-d30b70b5bed3" containerName="extract-content" Dec 01 22:10:53 crc kubenswrapper[4857]: E1201 22:10:53.410702 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3106c12-2d34-416c-ae4f-d30b70b5bed3" containerName="extract-utilities" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.410709 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3106c12-2d34-416c-ae4f-d30b70b5bed3" containerName="extract-utilities" Dec 01 22:10:53 crc kubenswrapper[4857]: E1201 22:10:53.410732 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f71f02cd-2132-437a-b18c-e3513eb0710c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.410741 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f71f02cd-2132-437a-b18c-e3513eb0710c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 22:10:53 crc kubenswrapper[4857]: E1201 22:10:53.410765 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3106c12-2d34-416c-ae4f-d30b70b5bed3" containerName="registry-server" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.410771 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3106c12-2d34-416c-ae4f-d30b70b5bed3" containerName="registry-server" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.410934 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3106c12-2d34-416c-ae4f-d30b70b5bed3" containerName="registry-server" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.410955 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f71f02cd-2132-437a-b18c-e3513eb0710c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.411603 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.416660 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.417214 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.417356 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.417441 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.417574 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.432676 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk"] Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.559600 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.559680 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6640e334-5d6f-441b-b213-985398e67768-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.559723 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wqcj\" (UniqueName: \"kubernetes.io/projected/6640e334-5d6f-441b-b213-985398e67768-kube-api-access-8wqcj\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.559838 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.559878 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.661531 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6640e334-5d6f-441b-b213-985398e67768-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.661591 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wqcj\" (UniqueName: \"kubernetes.io/projected/6640e334-5d6f-441b-b213-985398e67768-kube-api-access-8wqcj\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.661726 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.661753 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.661813 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.662929 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6640e334-5d6f-441b-b213-985398e67768-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.667556 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.667695 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.667953 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.681780 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wqcj\" (UniqueName: \"kubernetes.io/projected/6640e334-5d6f-441b-b213-985398e67768-kube-api-access-8wqcj\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-h6nrk\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:53 crc kubenswrapper[4857]: I1201 22:10:53.735773 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:10:54 crc kubenswrapper[4857]: I1201 22:10:54.314625 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk"] Dec 01 22:10:55 crc kubenswrapper[4857]: I1201 22:10:55.202211 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" event={"ID":"6640e334-5d6f-441b-b213-985398e67768","Type":"ContainerStarted","Data":"9f425760d39430234132cc768d48932e069602e61425b1a6892a58b046de20d9"} Dec 01 22:10:56 crc kubenswrapper[4857]: I1201 22:10:56.212778 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" event={"ID":"6640e334-5d6f-441b-b213-985398e67768","Type":"ContainerStarted","Data":"d206089a7e969b21a7d0c894e2b1951effc8c3bf51c4db2eaf38c21b110be8e8"} Dec 01 22:10:56 crc kubenswrapper[4857]: I1201 22:10:56.239086 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" podStartSLOduration=2.558415009 podStartE2EDuration="3.239063492s" podCreationTimestamp="2025-12-01 22:10:53 +0000 UTC" firstStartedPulling="2025-12-01 22:10:54.324285655 +0000 UTC m=+2172.814347982" lastFinishedPulling="2025-12-01 22:10:55.004934138 +0000 UTC m=+2173.494996465" observedRunningTime="2025-12-01 22:10:56.231270936 +0000 UTC m=+2174.721333623" watchObservedRunningTime="2025-12-01 22:10:56.239063492 +0000 UTC m=+2174.729125809" Dec 01 22:10:57 crc kubenswrapper[4857]: I1201 22:10:57.805993 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:10:57 crc kubenswrapper[4857]: I1201 22:10:57.806111 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:11:27 crc kubenswrapper[4857]: I1201 22:11:27.807576 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:11:27 crc kubenswrapper[4857]: I1201 22:11:27.808427 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:11:27 crc kubenswrapper[4857]: I1201 22:11:27.808486 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 22:11:27 crc kubenswrapper[4857]: I1201 22:11:27.809550 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d86df5e45e1b4e70e1cb13ada553f4b40c37fb697e38ddf8687ffc69d2758e7f"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 22:11:27 crc kubenswrapper[4857]: I1201 22:11:27.809611 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://d86df5e45e1b4e70e1cb13ada553f4b40c37fb697e38ddf8687ffc69d2758e7f" gracePeriod=600 Dec 01 22:11:28 crc kubenswrapper[4857]: I1201 22:11:28.611506 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="d86df5e45e1b4e70e1cb13ada553f4b40c37fb697e38ddf8687ffc69d2758e7f" exitCode=0 Dec 01 22:11:28 crc kubenswrapper[4857]: I1201 22:11:28.611581 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"d86df5e45e1b4e70e1cb13ada553f4b40c37fb697e38ddf8687ffc69d2758e7f"} Dec 01 22:11:28 crc kubenswrapper[4857]: I1201 22:11:28.612452 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159"} Dec 01 22:11:28 crc kubenswrapper[4857]: I1201 22:11:28.612481 4857 scope.go:117] "RemoveContainer" containerID="335c5fd119b007c965d65ee1a7857cbd74d6506734678ea4f96c92d1194ccd35" Dec 01 22:12:09 crc kubenswrapper[4857]: I1201 22:12:09.100402 4857 generic.go:334] "Generic (PLEG): container finished" podID="6640e334-5d6f-441b-b213-985398e67768" containerID="d206089a7e969b21a7d0c894e2b1951effc8c3bf51c4db2eaf38c21b110be8e8" exitCode=0 Dec 01 22:12:09 crc kubenswrapper[4857]: I1201 22:12:09.100532 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" event={"ID":"6640e334-5d6f-441b-b213-985398e67768","Type":"ContainerDied","Data":"d206089a7e969b21a7d0c894e2b1951effc8c3bf51c4db2eaf38c21b110be8e8"} Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.684555 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.779799 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wqcj\" (UniqueName: \"kubernetes.io/projected/6640e334-5d6f-441b-b213-985398e67768-kube-api-access-8wqcj\") pod \"6640e334-5d6f-441b-b213-985398e67768\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.780113 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-inventory\") pod \"6640e334-5d6f-441b-b213-985398e67768\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.780149 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-ovn-combined-ca-bundle\") pod \"6640e334-5d6f-441b-b213-985398e67768\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.780388 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6640e334-5d6f-441b-b213-985398e67768-ovncontroller-config-0\") pod \"6640e334-5d6f-441b-b213-985398e67768\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.780468 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-ssh-key\") pod \"6640e334-5d6f-441b-b213-985398e67768\" (UID: \"6640e334-5d6f-441b-b213-985398e67768\") " Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.785958 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6640e334-5d6f-441b-b213-985398e67768" (UID: "6640e334-5d6f-441b-b213-985398e67768"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.786046 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6640e334-5d6f-441b-b213-985398e67768-kube-api-access-8wqcj" (OuterVolumeSpecName: "kube-api-access-8wqcj") pod "6640e334-5d6f-441b-b213-985398e67768" (UID: "6640e334-5d6f-441b-b213-985398e67768"). InnerVolumeSpecName "kube-api-access-8wqcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.807354 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6640e334-5d6f-441b-b213-985398e67768" (UID: "6640e334-5d6f-441b-b213-985398e67768"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.812700 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6640e334-5d6f-441b-b213-985398e67768-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "6640e334-5d6f-441b-b213-985398e67768" (UID: "6640e334-5d6f-441b-b213-985398e67768"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.816163 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-inventory" (OuterVolumeSpecName: "inventory") pod "6640e334-5d6f-441b-b213-985398e67768" (UID: "6640e334-5d6f-441b-b213-985398e67768"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.883921 4857 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6640e334-5d6f-441b-b213-985398e67768-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.883947 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.883956 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wqcj\" (UniqueName: \"kubernetes.io/projected/6640e334-5d6f-441b-b213-985398e67768-kube-api-access-8wqcj\") on node \"crc\" DevicePath \"\"" Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.883966 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:12:10 crc kubenswrapper[4857]: I1201 22:12:10.883973 4857 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6640e334-5d6f-441b-b213-985398e67768-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.131154 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" event={"ID":"6640e334-5d6f-441b-b213-985398e67768","Type":"ContainerDied","Data":"9f425760d39430234132cc768d48932e069602e61425b1a6892a58b046de20d9"} Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.131195 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-h6nrk" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.131208 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f425760d39430234132cc768d48932e069602e61425b1a6892a58b046de20d9" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.250172 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x"] Dec 01 22:12:11 crc kubenswrapper[4857]: E1201 22:12:11.250539 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6640e334-5d6f-441b-b213-985398e67768" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.250556 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="6640e334-5d6f-441b-b213-985398e67768" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.250755 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="6640e334-5d6f-441b-b213-985398e67768" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.251461 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.258022 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.258617 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.258785 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.258981 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.258634 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.259641 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.272191 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x"] Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.323547 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.323747 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.323835 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.324041 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8ztx\" (UniqueName: \"kubernetes.io/projected/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-kube-api-access-n8ztx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.324121 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.324155 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.424913 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.424971 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.425004 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.425076 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8ztx\" (UniqueName: \"kubernetes.io/projected/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-kube-api-access-n8ztx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.425112 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.425130 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.430764 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.431539 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.432907 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.432966 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.436708 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.440841 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8ztx\" (UniqueName: \"kubernetes.io/projected/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-kube-api-access-n8ztx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:11 crc kubenswrapper[4857]: I1201 22:12:11.636584 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:12:12 crc kubenswrapper[4857]: W1201 22:12:12.223636 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60f230d9_9293_4ab7_a2c5_5a9e3c5dd623.slice/crio-fa9c03076bce664e548b25edc320d199a142237237f9e5c0485ff4628b6e9bfe WatchSource:0}: Error finding container fa9c03076bce664e548b25edc320d199a142237237f9e5c0485ff4628b6e9bfe: Status 404 returned error can't find the container with id fa9c03076bce664e548b25edc320d199a142237237f9e5c0485ff4628b6e9bfe Dec 01 22:12:12 crc kubenswrapper[4857]: I1201 22:12:12.237135 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x"] Dec 01 22:12:13 crc kubenswrapper[4857]: I1201 22:12:13.147798 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" event={"ID":"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623","Type":"ContainerStarted","Data":"feedf2bf1498a5b885c3616542771c63b8ef94b0e464f363565e1695542ddaa1"} Dec 01 22:12:13 crc kubenswrapper[4857]: I1201 22:12:13.148274 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" event={"ID":"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623","Type":"ContainerStarted","Data":"fa9c03076bce664e548b25edc320d199a142237237f9e5c0485ff4628b6e9bfe"} Dec 01 22:12:13 crc kubenswrapper[4857]: I1201 22:12:13.181425 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" podStartSLOduration=1.71392479 podStartE2EDuration="2.181397827s" podCreationTimestamp="2025-12-01 22:12:11 +0000 UTC" firstStartedPulling="2025-12-01 22:12:12.226869797 +0000 UTC m=+2250.716932124" lastFinishedPulling="2025-12-01 22:12:12.694342804 +0000 UTC m=+2251.184405161" observedRunningTime="2025-12-01 22:12:13.169601847 +0000 UTC m=+2251.659664214" watchObservedRunningTime="2025-12-01 22:12:13.181397827 +0000 UTC m=+2251.671460194" Dec 01 22:12:37 crc kubenswrapper[4857]: I1201 22:12:37.927397 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sfccx"] Dec 01 22:12:37 crc kubenswrapper[4857]: I1201 22:12:37.940910 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sfccx"] Dec 01 22:12:37 crc kubenswrapper[4857]: I1201 22:12:37.943141 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:38 crc kubenswrapper[4857]: I1201 22:12:38.086540 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15e40d3-01c3-48c8-af0e-420f694b0f10-catalog-content\") pod \"certified-operators-sfccx\" (UID: \"a15e40d3-01c3-48c8-af0e-420f694b0f10\") " pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:38 crc kubenswrapper[4857]: I1201 22:12:38.087166 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tmnz\" (UniqueName: \"kubernetes.io/projected/a15e40d3-01c3-48c8-af0e-420f694b0f10-kube-api-access-8tmnz\") pod \"certified-operators-sfccx\" (UID: \"a15e40d3-01c3-48c8-af0e-420f694b0f10\") " pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:38 crc kubenswrapper[4857]: I1201 22:12:38.087462 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15e40d3-01c3-48c8-af0e-420f694b0f10-utilities\") pod \"certified-operators-sfccx\" (UID: \"a15e40d3-01c3-48c8-af0e-420f694b0f10\") " pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:38 crc kubenswrapper[4857]: I1201 22:12:38.189092 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tmnz\" (UniqueName: \"kubernetes.io/projected/a15e40d3-01c3-48c8-af0e-420f694b0f10-kube-api-access-8tmnz\") pod \"certified-operators-sfccx\" (UID: \"a15e40d3-01c3-48c8-af0e-420f694b0f10\") " pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:38 crc kubenswrapper[4857]: I1201 22:12:38.189577 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15e40d3-01c3-48c8-af0e-420f694b0f10-utilities\") pod \"certified-operators-sfccx\" (UID: \"a15e40d3-01c3-48c8-af0e-420f694b0f10\") " pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:38 crc kubenswrapper[4857]: I1201 22:12:38.189756 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15e40d3-01c3-48c8-af0e-420f694b0f10-catalog-content\") pod \"certified-operators-sfccx\" (UID: \"a15e40d3-01c3-48c8-af0e-420f694b0f10\") " pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:38 crc kubenswrapper[4857]: I1201 22:12:38.190500 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15e40d3-01c3-48c8-af0e-420f694b0f10-catalog-content\") pod \"certified-operators-sfccx\" (UID: \"a15e40d3-01c3-48c8-af0e-420f694b0f10\") " pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:38 crc kubenswrapper[4857]: I1201 22:12:38.190532 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15e40d3-01c3-48c8-af0e-420f694b0f10-utilities\") pod \"certified-operators-sfccx\" (UID: \"a15e40d3-01c3-48c8-af0e-420f694b0f10\") " pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:38 crc kubenswrapper[4857]: I1201 22:12:38.216391 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tmnz\" (UniqueName: \"kubernetes.io/projected/a15e40d3-01c3-48c8-af0e-420f694b0f10-kube-api-access-8tmnz\") pod \"certified-operators-sfccx\" (UID: \"a15e40d3-01c3-48c8-af0e-420f694b0f10\") " pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:38 crc kubenswrapper[4857]: I1201 22:12:38.283547 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:38 crc kubenswrapper[4857]: I1201 22:12:38.823135 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sfccx"] Dec 01 22:12:39 crc kubenswrapper[4857]: I1201 22:12:39.460227 4857 generic.go:334] "Generic (PLEG): container finished" podID="a15e40d3-01c3-48c8-af0e-420f694b0f10" containerID="3d2c1afac98ab75516648772e17c701044bca3e7a932c2c1e1fd81f84ec4aadd" exitCode=0 Dec 01 22:12:39 crc kubenswrapper[4857]: I1201 22:12:39.460307 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sfccx" event={"ID":"a15e40d3-01c3-48c8-af0e-420f694b0f10","Type":"ContainerDied","Data":"3d2c1afac98ab75516648772e17c701044bca3e7a932c2c1e1fd81f84ec4aadd"} Dec 01 22:12:39 crc kubenswrapper[4857]: I1201 22:12:39.460544 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sfccx" event={"ID":"a15e40d3-01c3-48c8-af0e-420f694b0f10","Type":"ContainerStarted","Data":"47f0dd8e85c01f2d2da86774e455b75999c6894fddd67544a0b7540e67a9e190"} Dec 01 22:12:40 crc kubenswrapper[4857]: I1201 22:12:40.476331 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sfccx" event={"ID":"a15e40d3-01c3-48c8-af0e-420f694b0f10","Type":"ContainerStarted","Data":"e0a338e4c614bac40748915c5d00e00bdd5f766bf3165c5c3a67d3bc8b2214a8"} Dec 01 22:12:41 crc kubenswrapper[4857]: I1201 22:12:41.494083 4857 generic.go:334] "Generic (PLEG): container finished" podID="a15e40d3-01c3-48c8-af0e-420f694b0f10" containerID="e0a338e4c614bac40748915c5d00e00bdd5f766bf3165c5c3a67d3bc8b2214a8" exitCode=0 Dec 01 22:12:41 crc kubenswrapper[4857]: I1201 22:12:41.494183 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sfccx" event={"ID":"a15e40d3-01c3-48c8-af0e-420f694b0f10","Type":"ContainerDied","Data":"e0a338e4c614bac40748915c5d00e00bdd5f766bf3165c5c3a67d3bc8b2214a8"} Dec 01 22:12:42 crc kubenswrapper[4857]: I1201 22:12:42.508394 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sfccx" event={"ID":"a15e40d3-01c3-48c8-af0e-420f694b0f10","Type":"ContainerStarted","Data":"d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4"} Dec 01 22:12:42 crc kubenswrapper[4857]: I1201 22:12:42.543421 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sfccx" podStartSLOduration=3.031385977 podStartE2EDuration="5.543396925s" podCreationTimestamp="2025-12-01 22:12:37 +0000 UTC" firstStartedPulling="2025-12-01 22:12:39.462723812 +0000 UTC m=+2277.952786129" lastFinishedPulling="2025-12-01 22:12:41.97473472 +0000 UTC m=+2280.464797077" observedRunningTime="2025-12-01 22:12:42.538750881 +0000 UTC m=+2281.028813208" watchObservedRunningTime="2025-12-01 22:12:42.543396925 +0000 UTC m=+2281.033459282" Dec 01 22:12:48 crc kubenswrapper[4857]: I1201 22:12:48.284501 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:48 crc kubenswrapper[4857]: I1201 22:12:48.285302 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:48 crc kubenswrapper[4857]: I1201 22:12:48.346370 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:48 crc kubenswrapper[4857]: I1201 22:12:48.630161 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:48 crc kubenswrapper[4857]: I1201 22:12:48.696928 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sfccx"] Dec 01 22:12:50 crc kubenswrapper[4857]: I1201 22:12:50.604993 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sfccx" podUID="a15e40d3-01c3-48c8-af0e-420f694b0f10" containerName="registry-server" containerID="cri-o://d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4" gracePeriod=2 Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.127806 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.300159 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tmnz\" (UniqueName: \"kubernetes.io/projected/a15e40d3-01c3-48c8-af0e-420f694b0f10-kube-api-access-8tmnz\") pod \"a15e40d3-01c3-48c8-af0e-420f694b0f10\" (UID: \"a15e40d3-01c3-48c8-af0e-420f694b0f10\") " Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.300332 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15e40d3-01c3-48c8-af0e-420f694b0f10-catalog-content\") pod \"a15e40d3-01c3-48c8-af0e-420f694b0f10\" (UID: \"a15e40d3-01c3-48c8-af0e-420f694b0f10\") " Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.300407 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15e40d3-01c3-48c8-af0e-420f694b0f10-utilities\") pod \"a15e40d3-01c3-48c8-af0e-420f694b0f10\" (UID: \"a15e40d3-01c3-48c8-af0e-420f694b0f10\") " Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.302084 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a15e40d3-01c3-48c8-af0e-420f694b0f10-utilities" (OuterVolumeSpecName: "utilities") pod "a15e40d3-01c3-48c8-af0e-420f694b0f10" (UID: "a15e40d3-01c3-48c8-af0e-420f694b0f10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.317889 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a15e40d3-01c3-48c8-af0e-420f694b0f10-kube-api-access-8tmnz" (OuterVolumeSpecName: "kube-api-access-8tmnz") pod "a15e40d3-01c3-48c8-af0e-420f694b0f10" (UID: "a15e40d3-01c3-48c8-af0e-420f694b0f10"). InnerVolumeSpecName "kube-api-access-8tmnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.402745 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tmnz\" (UniqueName: \"kubernetes.io/projected/a15e40d3-01c3-48c8-af0e-420f694b0f10-kube-api-access-8tmnz\") on node \"crc\" DevicePath \"\"" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.402783 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15e40d3-01c3-48c8-af0e-420f694b0f10-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.477288 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a15e40d3-01c3-48c8-af0e-420f694b0f10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a15e40d3-01c3-48c8-af0e-420f694b0f10" (UID: "a15e40d3-01c3-48c8-af0e-420f694b0f10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.505622 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15e40d3-01c3-48c8-af0e-420f694b0f10-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.621482 4857 generic.go:334] "Generic (PLEG): container finished" podID="a15e40d3-01c3-48c8-af0e-420f694b0f10" containerID="d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4" exitCode=0 Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.621540 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sfccx" event={"ID":"a15e40d3-01c3-48c8-af0e-420f694b0f10","Type":"ContainerDied","Data":"d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4"} Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.621584 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sfccx" event={"ID":"a15e40d3-01c3-48c8-af0e-420f694b0f10","Type":"ContainerDied","Data":"47f0dd8e85c01f2d2da86774e455b75999c6894fddd67544a0b7540e67a9e190"} Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.621591 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sfccx" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.621615 4857 scope.go:117] "RemoveContainer" containerID="d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.676249 4857 scope.go:117] "RemoveContainer" containerID="e0a338e4c614bac40748915c5d00e00bdd5f766bf3165c5c3a67d3bc8b2214a8" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.682914 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sfccx"] Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.695587 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sfccx"] Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.715283 4857 scope.go:117] "RemoveContainer" containerID="3d2c1afac98ab75516648772e17c701044bca3e7a932c2c1e1fd81f84ec4aadd" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.773982 4857 scope.go:117] "RemoveContainer" containerID="d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4" Dec 01 22:12:51 crc kubenswrapper[4857]: E1201 22:12:51.775607 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4\": container with ID starting with d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4 not found: ID does not exist" containerID="d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.775668 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4"} err="failed to get container status \"d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4\": rpc error: code = NotFound desc = could not find container \"d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4\": container with ID starting with d14245773924d7355533310b96db15b54e968e08b66b4675a93cd43834080ae4 not found: ID does not exist" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.775713 4857 scope.go:117] "RemoveContainer" containerID="e0a338e4c614bac40748915c5d00e00bdd5f766bf3165c5c3a67d3bc8b2214a8" Dec 01 22:12:51 crc kubenswrapper[4857]: E1201 22:12:51.776132 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0a338e4c614bac40748915c5d00e00bdd5f766bf3165c5c3a67d3bc8b2214a8\": container with ID starting with e0a338e4c614bac40748915c5d00e00bdd5f766bf3165c5c3a67d3bc8b2214a8 not found: ID does not exist" containerID="e0a338e4c614bac40748915c5d00e00bdd5f766bf3165c5c3a67d3bc8b2214a8" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.776172 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0a338e4c614bac40748915c5d00e00bdd5f766bf3165c5c3a67d3bc8b2214a8"} err="failed to get container status \"e0a338e4c614bac40748915c5d00e00bdd5f766bf3165c5c3a67d3bc8b2214a8\": rpc error: code = NotFound desc = could not find container \"e0a338e4c614bac40748915c5d00e00bdd5f766bf3165c5c3a67d3bc8b2214a8\": container with ID starting with e0a338e4c614bac40748915c5d00e00bdd5f766bf3165c5c3a67d3bc8b2214a8 not found: ID does not exist" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.776202 4857 scope.go:117] "RemoveContainer" containerID="3d2c1afac98ab75516648772e17c701044bca3e7a932c2c1e1fd81f84ec4aadd" Dec 01 22:12:51 crc kubenswrapper[4857]: E1201 22:12:51.776690 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d2c1afac98ab75516648772e17c701044bca3e7a932c2c1e1fd81f84ec4aadd\": container with ID starting with 3d2c1afac98ab75516648772e17c701044bca3e7a932c2c1e1fd81f84ec4aadd not found: ID does not exist" containerID="3d2c1afac98ab75516648772e17c701044bca3e7a932c2c1e1fd81f84ec4aadd" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.776757 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d2c1afac98ab75516648772e17c701044bca3e7a932c2c1e1fd81f84ec4aadd"} err="failed to get container status \"3d2c1afac98ab75516648772e17c701044bca3e7a932c2c1e1fd81f84ec4aadd\": rpc error: code = NotFound desc = could not find container \"3d2c1afac98ab75516648772e17c701044bca3e7a932c2c1e1fd81f84ec4aadd\": container with ID starting with 3d2c1afac98ab75516648772e17c701044bca3e7a932c2c1e1fd81f84ec4aadd not found: ID does not exist" Dec 01 22:12:51 crc kubenswrapper[4857]: I1201 22:12:51.848004 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a15e40d3-01c3-48c8-af0e-420f694b0f10" path="/var/lib/kubelet/pods/a15e40d3-01c3-48c8-af0e-420f694b0f10/volumes" Dec 01 22:13:09 crc kubenswrapper[4857]: I1201 22:13:09.865243 4857 generic.go:334] "Generic (PLEG): container finished" podID="60f230d9-9293-4ab7-a2c5-5a9e3c5dd623" containerID="feedf2bf1498a5b885c3616542771c63b8ef94b0e464f363565e1695542ddaa1" exitCode=0 Dec 01 22:13:09 crc kubenswrapper[4857]: I1201 22:13:09.865360 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" event={"ID":"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623","Type":"ContainerDied","Data":"feedf2bf1498a5b885c3616542771c63b8ef94b0e464f363565e1695542ddaa1"} Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.314767 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.380974 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-ssh-key\") pod \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.381069 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-inventory\") pod \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.381119 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-nova-metadata-neutron-config-0\") pod \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.381198 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8ztx\" (UniqueName: \"kubernetes.io/projected/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-kube-api-access-n8ztx\") pod \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.381303 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-neutron-metadata-combined-ca-bundle\") pod \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.381408 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-neutron-ovn-metadata-agent-neutron-config-0\") pod \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\" (UID: \"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623\") " Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.387297 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "60f230d9-9293-4ab7-a2c5-5a9e3c5dd623" (UID: "60f230d9-9293-4ab7-a2c5-5a9e3c5dd623"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.388603 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-kube-api-access-n8ztx" (OuterVolumeSpecName: "kube-api-access-n8ztx") pod "60f230d9-9293-4ab7-a2c5-5a9e3c5dd623" (UID: "60f230d9-9293-4ab7-a2c5-5a9e3c5dd623"). InnerVolumeSpecName "kube-api-access-n8ztx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.420198 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-inventory" (OuterVolumeSpecName: "inventory") pod "60f230d9-9293-4ab7-a2c5-5a9e3c5dd623" (UID: "60f230d9-9293-4ab7-a2c5-5a9e3c5dd623"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.423037 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "60f230d9-9293-4ab7-a2c5-5a9e3c5dd623" (UID: "60f230d9-9293-4ab7-a2c5-5a9e3c5dd623"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.423427 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "60f230d9-9293-4ab7-a2c5-5a9e3c5dd623" (UID: "60f230d9-9293-4ab7-a2c5-5a9e3c5dd623"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.445406 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "60f230d9-9293-4ab7-a2c5-5a9e3c5dd623" (UID: "60f230d9-9293-4ab7-a2c5-5a9e3c5dd623"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.483740 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8ztx\" (UniqueName: \"kubernetes.io/projected/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-kube-api-access-n8ztx\") on node \"crc\" DevicePath \"\"" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.484065 4857 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.484166 4857 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.484250 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.484345 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.484429 4857 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/60f230d9-9293-4ab7-a2c5-5a9e3c5dd623-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.897964 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" event={"ID":"60f230d9-9293-4ab7-a2c5-5a9e3c5dd623","Type":"ContainerDied","Data":"fa9c03076bce664e548b25edc320d199a142237237f9e5c0485ff4628b6e9bfe"} Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.898314 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa9c03076bce664e548b25edc320d199a142237237f9e5c0485ff4628b6e9bfe" Dec 01 22:13:11 crc kubenswrapper[4857]: I1201 22:13:11.898037 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.072604 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4"] Dec 01 22:13:12 crc kubenswrapper[4857]: E1201 22:13:12.073375 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a15e40d3-01c3-48c8-af0e-420f694b0f10" containerName="registry-server" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.073480 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a15e40d3-01c3-48c8-af0e-420f694b0f10" containerName="registry-server" Dec 01 22:13:12 crc kubenswrapper[4857]: E1201 22:13:12.073568 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a15e40d3-01c3-48c8-af0e-420f694b0f10" containerName="extract-utilities" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.073654 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a15e40d3-01c3-48c8-af0e-420f694b0f10" containerName="extract-utilities" Dec 01 22:13:12 crc kubenswrapper[4857]: E1201 22:13:12.073737 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a15e40d3-01c3-48c8-af0e-420f694b0f10" containerName="extract-content" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.073795 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a15e40d3-01c3-48c8-af0e-420f694b0f10" containerName="extract-content" Dec 01 22:13:12 crc kubenswrapper[4857]: E1201 22:13:12.073877 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60f230d9-9293-4ab7-a2c5-5a9e3c5dd623" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.073931 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="60f230d9-9293-4ab7-a2c5-5a9e3c5dd623" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.074182 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="60f230d9-9293-4ab7-a2c5-5a9e3c5dd623" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.074280 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="a15e40d3-01c3-48c8-af0e-420f694b0f10" containerName="registry-server" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.075076 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.078283 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.078605 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.078996 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.079331 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.079607 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.080448 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4"] Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.201443 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.201577 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcrmj\" (UniqueName: \"kubernetes.io/projected/afe24796-7855-411d-b62a-a4e5fefc12c6-kube-api-access-lcrmj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.201639 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.201665 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.201687 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.304001 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcrmj\" (UniqueName: \"kubernetes.io/projected/afe24796-7855-411d-b62a-a4e5fefc12c6-kube-api-access-lcrmj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.304204 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.304274 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.304319 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.304517 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.309405 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.310455 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.310823 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.311426 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.328240 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcrmj\" (UniqueName: \"kubernetes.io/projected/afe24796-7855-411d-b62a-a4e5fefc12c6-kube-api-access-lcrmj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:12 crc kubenswrapper[4857]: I1201 22:13:12.391798 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:13:13 crc kubenswrapper[4857]: I1201 22:13:13.057411 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4"] Dec 01 22:13:13 crc kubenswrapper[4857]: I1201 22:13:13.922709 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" event={"ID":"afe24796-7855-411d-b62a-a4e5fefc12c6","Type":"ContainerStarted","Data":"dd5edb846a076fb0221a0162951747957569548233a2d6f927c1c45b91bf1bf5"} Dec 01 22:13:13 crc kubenswrapper[4857]: I1201 22:13:13.923354 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" event={"ID":"afe24796-7855-411d-b62a-a4e5fefc12c6","Type":"ContainerStarted","Data":"adea35d4ba781c1a20b5f42b540358f81c7af868d934e9b29d05eefbfa472758"} Dec 01 22:13:13 crc kubenswrapper[4857]: I1201 22:13:13.957169 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" podStartSLOduration=2.522313917 podStartE2EDuration="2.957117301s" podCreationTimestamp="2025-12-01 22:13:11 +0000 UTC" firstStartedPulling="2025-12-01 22:13:13.054707742 +0000 UTC m=+2311.544770069" lastFinishedPulling="2025-12-01 22:13:13.489511126 +0000 UTC m=+2311.979573453" observedRunningTime="2025-12-01 22:13:13.943849967 +0000 UTC m=+2312.433912324" watchObservedRunningTime="2025-12-01 22:13:13.957117301 +0000 UTC m=+2312.447179648" Dec 01 22:13:57 crc kubenswrapper[4857]: I1201 22:13:57.806630 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:13:57 crc kubenswrapper[4857]: I1201 22:13:57.807243 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:14:27 crc kubenswrapper[4857]: I1201 22:14:27.805966 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:14:27 crc kubenswrapper[4857]: I1201 22:14:27.806981 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:14:57 crc kubenswrapper[4857]: I1201 22:14:57.806635 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:14:57 crc kubenswrapper[4857]: I1201 22:14:57.807232 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:14:57 crc kubenswrapper[4857]: I1201 22:14:57.807314 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 22:14:57 crc kubenswrapper[4857]: I1201 22:14:57.808223 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 22:14:57 crc kubenswrapper[4857]: I1201 22:14:57.808320 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" gracePeriod=600 Dec 01 22:14:57 crc kubenswrapper[4857]: E1201 22:14:57.951845 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:14:58 crc kubenswrapper[4857]: I1201 22:14:58.191746 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" exitCode=0 Dec 01 22:14:58 crc kubenswrapper[4857]: I1201 22:14:58.191804 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159"} Dec 01 22:14:58 crc kubenswrapper[4857]: I1201 22:14:58.191856 4857 scope.go:117] "RemoveContainer" containerID="d86df5e45e1b4e70e1cb13ada553f4b40c37fb697e38ddf8687ffc69d2758e7f" Dec 01 22:14:58 crc kubenswrapper[4857]: I1201 22:14:58.193118 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:14:58 crc kubenswrapper[4857]: E1201 22:14:58.193768 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.184500 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l"] Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.187214 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.191055 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.191103 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.193558 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l"] Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.279925 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6pq7\" (UniqueName: \"kubernetes.io/projected/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-kube-api-access-z6pq7\") pod \"collect-profiles-29410455-sct6l\" (UID: \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.280133 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-secret-volume\") pod \"collect-profiles-29410455-sct6l\" (UID: \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.280240 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-config-volume\") pod \"collect-profiles-29410455-sct6l\" (UID: \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.382646 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-config-volume\") pod \"collect-profiles-29410455-sct6l\" (UID: \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.382819 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6pq7\" (UniqueName: \"kubernetes.io/projected/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-kube-api-access-z6pq7\") pod \"collect-profiles-29410455-sct6l\" (UID: \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.382887 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-secret-volume\") pod \"collect-profiles-29410455-sct6l\" (UID: \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.384615 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-config-volume\") pod \"collect-profiles-29410455-sct6l\" (UID: \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.392206 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-secret-volume\") pod \"collect-profiles-29410455-sct6l\" (UID: \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.405551 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6pq7\" (UniqueName: \"kubernetes.io/projected/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-kube-api-access-z6pq7\") pod \"collect-profiles-29410455-sct6l\" (UID: \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:00 crc kubenswrapper[4857]: I1201 22:15:00.538552 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:01 crc kubenswrapper[4857]: I1201 22:15:01.099514 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l"] Dec 01 22:15:01 crc kubenswrapper[4857]: I1201 22:15:01.238673 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" event={"ID":"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a","Type":"ContainerStarted","Data":"995dc454544584aeee9c7cac51df685ab7401c2bc04a246539a6548484b0925d"} Dec 01 22:15:02 crc kubenswrapper[4857]: I1201 22:15:02.251416 4857 generic.go:334] "Generic (PLEG): container finished" podID="891266c7-ed7c-48c1-bbe9-2fba0bb9da3a" containerID="54d396b7c4ef79189f08065bb8aeb17cfc9fa75fd255d398195f68bb663f6d34" exitCode=0 Dec 01 22:15:02 crc kubenswrapper[4857]: I1201 22:15:02.251632 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" event={"ID":"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a","Type":"ContainerDied","Data":"54d396b7c4ef79189f08065bb8aeb17cfc9fa75fd255d398195f68bb663f6d34"} Dec 01 22:15:03 crc kubenswrapper[4857]: I1201 22:15:03.731573 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:03 crc kubenswrapper[4857]: I1201 22:15:03.754554 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-secret-volume\") pod \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\" (UID: \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\") " Dec 01 22:15:03 crc kubenswrapper[4857]: I1201 22:15:03.754649 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6pq7\" (UniqueName: \"kubernetes.io/projected/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-kube-api-access-z6pq7\") pod \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\" (UID: \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\") " Dec 01 22:15:03 crc kubenswrapper[4857]: I1201 22:15:03.754777 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-config-volume\") pod \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\" (UID: \"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a\") " Dec 01 22:15:03 crc kubenswrapper[4857]: I1201 22:15:03.756330 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-config-volume" (OuterVolumeSpecName: "config-volume") pod "891266c7-ed7c-48c1-bbe9-2fba0bb9da3a" (UID: "891266c7-ed7c-48c1-bbe9-2fba0bb9da3a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:15:03 crc kubenswrapper[4857]: I1201 22:15:03.761757 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-kube-api-access-z6pq7" (OuterVolumeSpecName: "kube-api-access-z6pq7") pod "891266c7-ed7c-48c1-bbe9-2fba0bb9da3a" (UID: "891266c7-ed7c-48c1-bbe9-2fba0bb9da3a"). InnerVolumeSpecName "kube-api-access-z6pq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:15:03 crc kubenswrapper[4857]: I1201 22:15:03.769965 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "891266c7-ed7c-48c1-bbe9-2fba0bb9da3a" (UID: "891266c7-ed7c-48c1-bbe9-2fba0bb9da3a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:15:03 crc kubenswrapper[4857]: I1201 22:15:03.857963 4857 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 22:15:03 crc kubenswrapper[4857]: I1201 22:15:03.858103 4857 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 22:15:03 crc kubenswrapper[4857]: I1201 22:15:03.858129 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6pq7\" (UniqueName: \"kubernetes.io/projected/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a-kube-api-access-z6pq7\") on node \"crc\" DevicePath \"\"" Dec 01 22:15:04 crc kubenswrapper[4857]: I1201 22:15:04.278160 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" event={"ID":"891266c7-ed7c-48c1-bbe9-2fba0bb9da3a","Type":"ContainerDied","Data":"995dc454544584aeee9c7cac51df685ab7401c2bc04a246539a6548484b0925d"} Dec 01 22:15:04 crc kubenswrapper[4857]: I1201 22:15:04.278196 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l" Dec 01 22:15:04 crc kubenswrapper[4857]: I1201 22:15:04.278224 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="995dc454544584aeee9c7cac51df685ab7401c2bc04a246539a6548484b0925d" Dec 01 22:15:04 crc kubenswrapper[4857]: I1201 22:15:04.821625 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb"] Dec 01 22:15:04 crc kubenswrapper[4857]: I1201 22:15:04.832299 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410410-9gpfb"] Dec 01 22:15:05 crc kubenswrapper[4857]: I1201 22:15:05.849236 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c8375eb-1803-475e-b897-9307d7bc29ee" path="/var/lib/kubelet/pods/6c8375eb-1803-475e-b897-9307d7bc29ee/volumes" Dec 01 22:15:13 crc kubenswrapper[4857]: I1201 22:15:13.835320 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:15:13 crc kubenswrapper[4857]: E1201 22:15:13.836357 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:15:22 crc kubenswrapper[4857]: I1201 22:15:22.631134 4857 scope.go:117] "RemoveContainer" containerID="fdc6a1bc8c1714b9add47cbd440bbc801690cba6bde6e6b375970539e22fdff7" Dec 01 22:15:26 crc kubenswrapper[4857]: I1201 22:15:26.835103 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:15:26 crc kubenswrapper[4857]: E1201 22:15:26.836233 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:15:41 crc kubenswrapper[4857]: I1201 22:15:41.846206 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:15:41 crc kubenswrapper[4857]: E1201 22:15:41.847106 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:15:53 crc kubenswrapper[4857]: I1201 22:15:53.835090 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:15:53 crc kubenswrapper[4857]: E1201 22:15:53.835960 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:16:04 crc kubenswrapper[4857]: I1201 22:16:04.836246 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:16:04 crc kubenswrapper[4857]: E1201 22:16:04.837293 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:16:15 crc kubenswrapper[4857]: I1201 22:16:15.835593 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:16:15 crc kubenswrapper[4857]: E1201 22:16:15.836516 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:16:29 crc kubenswrapper[4857]: I1201 22:16:29.836212 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:16:29 crc kubenswrapper[4857]: E1201 22:16:29.837591 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:16:42 crc kubenswrapper[4857]: I1201 22:16:42.835571 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:16:42 crc kubenswrapper[4857]: E1201 22:16:42.836984 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:16:54 crc kubenswrapper[4857]: I1201 22:16:54.836113 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:16:54 crc kubenswrapper[4857]: E1201 22:16:54.837246 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:17:08 crc kubenswrapper[4857]: I1201 22:17:08.834348 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:17:08 crc kubenswrapper[4857]: E1201 22:17:08.835246 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:17:23 crc kubenswrapper[4857]: I1201 22:17:23.835064 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:17:23 crc kubenswrapper[4857]: E1201 22:17:23.836149 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:17:37 crc kubenswrapper[4857]: I1201 22:17:37.836673 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:17:37 crc kubenswrapper[4857]: E1201 22:17:37.837754 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:17:50 crc kubenswrapper[4857]: I1201 22:17:50.835722 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:17:50 crc kubenswrapper[4857]: E1201 22:17:50.838300 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:17:55 crc kubenswrapper[4857]: I1201 22:17:55.410642 4857 generic.go:334] "Generic (PLEG): container finished" podID="afe24796-7855-411d-b62a-a4e5fefc12c6" containerID="dd5edb846a076fb0221a0162951747957569548233a2d6f927c1c45b91bf1bf5" exitCode=0 Dec 01 22:17:55 crc kubenswrapper[4857]: I1201 22:17:55.411210 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" event={"ID":"afe24796-7855-411d-b62a-a4e5fefc12c6","Type":"ContainerDied","Data":"dd5edb846a076fb0221a0162951747957569548233a2d6f927c1c45b91bf1bf5"} Dec 01 22:17:56 crc kubenswrapper[4857]: I1201 22:17:56.958398 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.067875 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-inventory\") pod \"afe24796-7855-411d-b62a-a4e5fefc12c6\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.067931 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-libvirt-combined-ca-bundle\") pod \"afe24796-7855-411d-b62a-a4e5fefc12c6\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.067970 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcrmj\" (UniqueName: \"kubernetes.io/projected/afe24796-7855-411d-b62a-a4e5fefc12c6-kube-api-access-lcrmj\") pod \"afe24796-7855-411d-b62a-a4e5fefc12c6\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.068021 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-ssh-key\") pod \"afe24796-7855-411d-b62a-a4e5fefc12c6\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.068062 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-libvirt-secret-0\") pod \"afe24796-7855-411d-b62a-a4e5fefc12c6\" (UID: \"afe24796-7855-411d-b62a-a4e5fefc12c6\") " Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.078140 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "afe24796-7855-411d-b62a-a4e5fefc12c6" (UID: "afe24796-7855-411d-b62a-a4e5fefc12c6"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.080910 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afe24796-7855-411d-b62a-a4e5fefc12c6-kube-api-access-lcrmj" (OuterVolumeSpecName: "kube-api-access-lcrmj") pod "afe24796-7855-411d-b62a-a4e5fefc12c6" (UID: "afe24796-7855-411d-b62a-a4e5fefc12c6"). InnerVolumeSpecName "kube-api-access-lcrmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.103195 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "afe24796-7855-411d-b62a-a4e5fefc12c6" (UID: "afe24796-7855-411d-b62a-a4e5fefc12c6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.113489 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-inventory" (OuterVolumeSpecName: "inventory") pod "afe24796-7855-411d-b62a-a4e5fefc12c6" (UID: "afe24796-7855-411d-b62a-a4e5fefc12c6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.123219 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "afe24796-7855-411d-b62a-a4e5fefc12c6" (UID: "afe24796-7855-411d-b62a-a4e5fefc12c6"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.171562 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.171620 4857 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.171644 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcrmj\" (UniqueName: \"kubernetes.io/projected/afe24796-7855-411d-b62a-a4e5fefc12c6-kube-api-access-lcrmj\") on node \"crc\" DevicePath \"\"" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.171660 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.171684 4857 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/afe24796-7855-411d-b62a-a4e5fefc12c6-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.435613 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" event={"ID":"afe24796-7855-411d-b62a-a4e5fefc12c6","Type":"ContainerDied","Data":"adea35d4ba781c1a20b5f42b540358f81c7af868d934e9b29d05eefbfa472758"} Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.435652 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adea35d4ba781c1a20b5f42b540358f81c7af868d934e9b29d05eefbfa472758" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.435735 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.580720 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc"] Dec 01 22:17:57 crc kubenswrapper[4857]: E1201 22:17:57.581164 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe24796-7855-411d-b62a-a4e5fefc12c6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.581180 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe24796-7855-411d-b62a-a4e5fefc12c6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 22:17:57 crc kubenswrapper[4857]: E1201 22:17:57.581193 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="891266c7-ed7c-48c1-bbe9-2fba0bb9da3a" containerName="collect-profiles" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.581198 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="891266c7-ed7c-48c1-bbe9-2fba0bb9da3a" containerName="collect-profiles" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.581398 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="891266c7-ed7c-48c1-bbe9-2fba0bb9da3a" containerName="collect-profiles" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.581414 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="afe24796-7855-411d-b62a-a4e5fefc12c6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.582144 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.586198 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.586739 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.588266 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.588322 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.588502 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.588518 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.588748 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.610595 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc"] Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.695165 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.695219 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.695251 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.695281 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krxb4\" (UniqueName: \"kubernetes.io/projected/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-kube-api-access-krxb4\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.695309 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.695351 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.695384 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.695424 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.695442 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.796026 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.796093 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.796130 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.796147 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.796205 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.796229 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.796252 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.796273 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krxb4\" (UniqueName: \"kubernetes.io/projected/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-kube-api-access-krxb4\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.796301 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.797441 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.800257 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.800398 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.800459 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.800685 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.802556 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.802674 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.803003 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.813292 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krxb4\" (UniqueName: \"kubernetes.io/projected/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-kube-api-access-krxb4\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l7ltc\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:57 crc kubenswrapper[4857]: I1201 22:17:57.919015 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:17:58 crc kubenswrapper[4857]: I1201 22:17:58.527810 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc"] Dec 01 22:17:58 crc kubenswrapper[4857]: W1201 22:17:58.528491 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b8d39f4_2804_4fda_bec9_0cd4c2401f42.slice/crio-db5f9059414aae02ff33e6b32db44d0ae706c6ceadfb358182cd4af16a5f7cde WatchSource:0}: Error finding container db5f9059414aae02ff33e6b32db44d0ae706c6ceadfb358182cd4af16a5f7cde: Status 404 returned error can't find the container with id db5f9059414aae02ff33e6b32db44d0ae706c6ceadfb358182cd4af16a5f7cde Dec 01 22:17:58 crc kubenswrapper[4857]: I1201 22:17:58.532861 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 22:17:59 crc kubenswrapper[4857]: I1201 22:17:59.458493 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" event={"ID":"3b8d39f4-2804-4fda-bec9-0cd4c2401f42","Type":"ContainerStarted","Data":"db5f9059414aae02ff33e6b32db44d0ae706c6ceadfb358182cd4af16a5f7cde"} Dec 01 22:18:00 crc kubenswrapper[4857]: I1201 22:18:00.472809 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" event={"ID":"3b8d39f4-2804-4fda-bec9-0cd4c2401f42","Type":"ContainerStarted","Data":"d1622483f2911e92330e1f5a81f0bcf6e957c9c14981300099800e8d58a2467a"} Dec 01 22:18:00 crc kubenswrapper[4857]: I1201 22:18:00.506570 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" podStartSLOduration=2.715815611 podStartE2EDuration="3.506545007s" podCreationTimestamp="2025-12-01 22:17:57 +0000 UTC" firstStartedPulling="2025-12-01 22:17:58.532351014 +0000 UTC m=+2597.022413341" lastFinishedPulling="2025-12-01 22:17:59.32308038 +0000 UTC m=+2597.813142737" observedRunningTime="2025-12-01 22:18:00.49439502 +0000 UTC m=+2598.984457367" watchObservedRunningTime="2025-12-01 22:18:00.506545007 +0000 UTC m=+2598.996607354" Dec 01 22:18:01 crc kubenswrapper[4857]: I1201 22:18:01.855782 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:18:01 crc kubenswrapper[4857]: E1201 22:18:01.856548 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:18:13 crc kubenswrapper[4857]: I1201 22:18:13.834828 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:18:13 crc kubenswrapper[4857]: E1201 22:18:13.835946 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:18:25 crc kubenswrapper[4857]: I1201 22:18:25.836095 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:18:25 crc kubenswrapper[4857]: E1201 22:18:25.837001 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:18:36 crc kubenswrapper[4857]: I1201 22:18:36.836165 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:18:36 crc kubenswrapper[4857]: E1201 22:18:36.837499 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:18:50 crc kubenswrapper[4857]: I1201 22:18:50.836686 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:18:50 crc kubenswrapper[4857]: E1201 22:18:50.839206 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.052731 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6ffhh"] Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.057223 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.074371 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6ffhh"] Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.079175 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcd98f11-32b4-4adc-81c2-28402405c9a9-catalog-content\") pod \"redhat-operators-6ffhh\" (UID: \"dcd98f11-32b4-4adc-81c2-28402405c9a9\") " pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.079430 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg42q\" (UniqueName: \"kubernetes.io/projected/dcd98f11-32b4-4adc-81c2-28402405c9a9-kube-api-access-wg42q\") pod \"redhat-operators-6ffhh\" (UID: \"dcd98f11-32b4-4adc-81c2-28402405c9a9\") " pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.079482 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcd98f11-32b4-4adc-81c2-28402405c9a9-utilities\") pod \"redhat-operators-6ffhh\" (UID: \"dcd98f11-32b4-4adc-81c2-28402405c9a9\") " pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.182635 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcd98f11-32b4-4adc-81c2-28402405c9a9-catalog-content\") pod \"redhat-operators-6ffhh\" (UID: \"dcd98f11-32b4-4adc-81c2-28402405c9a9\") " pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.183395 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg42q\" (UniqueName: \"kubernetes.io/projected/dcd98f11-32b4-4adc-81c2-28402405c9a9-kube-api-access-wg42q\") pod \"redhat-operators-6ffhh\" (UID: \"dcd98f11-32b4-4adc-81c2-28402405c9a9\") " pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.183735 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcd98f11-32b4-4adc-81c2-28402405c9a9-utilities\") pod \"redhat-operators-6ffhh\" (UID: \"dcd98f11-32b4-4adc-81c2-28402405c9a9\") " pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.183592 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcd98f11-32b4-4adc-81c2-28402405c9a9-catalog-content\") pod \"redhat-operators-6ffhh\" (UID: \"dcd98f11-32b4-4adc-81c2-28402405c9a9\") " pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.184728 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcd98f11-32b4-4adc-81c2-28402405c9a9-utilities\") pod \"redhat-operators-6ffhh\" (UID: \"dcd98f11-32b4-4adc-81c2-28402405c9a9\") " pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.222801 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg42q\" (UniqueName: \"kubernetes.io/projected/dcd98f11-32b4-4adc-81c2-28402405c9a9-kube-api-access-wg42q\") pod \"redhat-operators-6ffhh\" (UID: \"dcd98f11-32b4-4adc-81c2-28402405c9a9\") " pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:18:55 crc kubenswrapper[4857]: I1201 22:18:55.435369 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:18:56 crc kubenswrapper[4857]: I1201 22:18:56.005224 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6ffhh"] Dec 01 22:18:56 crc kubenswrapper[4857]: I1201 22:18:56.153627 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ffhh" event={"ID":"dcd98f11-32b4-4adc-81c2-28402405c9a9","Type":"ContainerStarted","Data":"c3474eb4094bf0d37cb4dbeaa8df4894512eaba7946fc556b27c464902273933"} Dec 01 22:18:57 crc kubenswrapper[4857]: I1201 22:18:57.172316 4857 generic.go:334] "Generic (PLEG): container finished" podID="dcd98f11-32b4-4adc-81c2-28402405c9a9" containerID="cd60d9229247d7dd66dc7687b4a18772c0b22a666f1d15d9e49246a35afa4397" exitCode=0 Dec 01 22:18:57 crc kubenswrapper[4857]: I1201 22:18:57.172391 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ffhh" event={"ID":"dcd98f11-32b4-4adc-81c2-28402405c9a9","Type":"ContainerDied","Data":"cd60d9229247d7dd66dc7687b4a18772c0b22a666f1d15d9e49246a35afa4397"} Dec 01 22:18:59 crc kubenswrapper[4857]: I1201 22:18:59.202268 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ffhh" event={"ID":"dcd98f11-32b4-4adc-81c2-28402405c9a9","Type":"ContainerStarted","Data":"9875651beb0fb03342e8312ca9a686d6bbdaac46043f90c747db546c18f27426"} Dec 01 22:19:01 crc kubenswrapper[4857]: I1201 22:19:01.232214 4857 generic.go:334] "Generic (PLEG): container finished" podID="dcd98f11-32b4-4adc-81c2-28402405c9a9" containerID="9875651beb0fb03342e8312ca9a686d6bbdaac46043f90c747db546c18f27426" exitCode=0 Dec 01 22:19:01 crc kubenswrapper[4857]: I1201 22:19:01.232337 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ffhh" event={"ID":"dcd98f11-32b4-4adc-81c2-28402405c9a9","Type":"ContainerDied","Data":"9875651beb0fb03342e8312ca9a686d6bbdaac46043f90c747db546c18f27426"} Dec 01 22:19:03 crc kubenswrapper[4857]: I1201 22:19:03.258981 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ffhh" event={"ID":"dcd98f11-32b4-4adc-81c2-28402405c9a9","Type":"ContainerStarted","Data":"012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1"} Dec 01 22:19:03 crc kubenswrapper[4857]: I1201 22:19:03.283403 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6ffhh" podStartSLOduration=2.528262447 podStartE2EDuration="8.283385964s" podCreationTimestamp="2025-12-01 22:18:55 +0000 UTC" firstStartedPulling="2025-12-01 22:18:57.175907207 +0000 UTC m=+2655.665969564" lastFinishedPulling="2025-12-01 22:19:02.931030764 +0000 UTC m=+2661.421093081" observedRunningTime="2025-12-01 22:19:03.278029103 +0000 UTC m=+2661.768091420" watchObservedRunningTime="2025-12-01 22:19:03.283385964 +0000 UTC m=+2661.773448281" Dec 01 22:19:04 crc kubenswrapper[4857]: I1201 22:19:04.835363 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:19:04 crc kubenswrapper[4857]: E1201 22:19:04.836257 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:19:05 crc kubenswrapper[4857]: I1201 22:19:05.435828 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:19:05 crc kubenswrapper[4857]: I1201 22:19:05.436414 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:19:06 crc kubenswrapper[4857]: I1201 22:19:06.601142 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6ffhh" podUID="dcd98f11-32b4-4adc-81c2-28402405c9a9" containerName="registry-server" probeResult="failure" output=< Dec 01 22:19:06 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Dec 01 22:19:06 crc kubenswrapper[4857]: > Dec 01 22:19:15 crc kubenswrapper[4857]: I1201 22:19:15.519160 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:19:15 crc kubenswrapper[4857]: I1201 22:19:15.591390 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:19:15 crc kubenswrapper[4857]: I1201 22:19:15.768978 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6ffhh"] Dec 01 22:19:17 crc kubenswrapper[4857]: I1201 22:19:17.427183 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6ffhh" podUID="dcd98f11-32b4-4adc-81c2-28402405c9a9" containerName="registry-server" containerID="cri-o://012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1" gracePeriod=2 Dec 01 22:19:17 crc kubenswrapper[4857]: I1201 22:19:17.971298 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.096550 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcd98f11-32b4-4adc-81c2-28402405c9a9-utilities\") pod \"dcd98f11-32b4-4adc-81c2-28402405c9a9\" (UID: \"dcd98f11-32b4-4adc-81c2-28402405c9a9\") " Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.097738 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcd98f11-32b4-4adc-81c2-28402405c9a9-catalog-content\") pod \"dcd98f11-32b4-4adc-81c2-28402405c9a9\" (UID: \"dcd98f11-32b4-4adc-81c2-28402405c9a9\") " Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.123229 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg42q\" (UniqueName: \"kubernetes.io/projected/dcd98f11-32b4-4adc-81c2-28402405c9a9-kube-api-access-wg42q\") pod \"dcd98f11-32b4-4adc-81c2-28402405c9a9\" (UID: \"dcd98f11-32b4-4adc-81c2-28402405c9a9\") " Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.097689 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcd98f11-32b4-4adc-81c2-28402405c9a9-utilities" (OuterVolumeSpecName: "utilities") pod "dcd98f11-32b4-4adc-81c2-28402405c9a9" (UID: "dcd98f11-32b4-4adc-81c2-28402405c9a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.134273 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcd98f11-32b4-4adc-81c2-28402405c9a9-kube-api-access-wg42q" (OuterVolumeSpecName: "kube-api-access-wg42q") pod "dcd98f11-32b4-4adc-81c2-28402405c9a9" (UID: "dcd98f11-32b4-4adc-81c2-28402405c9a9"). InnerVolumeSpecName "kube-api-access-wg42q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.226986 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcd98f11-32b4-4adc-81c2-28402405c9a9-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.227258 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg42q\" (UniqueName: \"kubernetes.io/projected/dcd98f11-32b4-4adc-81c2-28402405c9a9-kube-api-access-wg42q\") on node \"crc\" DevicePath \"\"" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.260314 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcd98f11-32b4-4adc-81c2-28402405c9a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dcd98f11-32b4-4adc-81c2-28402405c9a9" (UID: "dcd98f11-32b4-4adc-81c2-28402405c9a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.328191 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcd98f11-32b4-4adc-81c2-28402405c9a9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.439203 4857 generic.go:334] "Generic (PLEG): container finished" podID="dcd98f11-32b4-4adc-81c2-28402405c9a9" containerID="012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1" exitCode=0 Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.439250 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ffhh" event={"ID":"dcd98f11-32b4-4adc-81c2-28402405c9a9","Type":"ContainerDied","Data":"012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1"} Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.439295 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ffhh" event={"ID":"dcd98f11-32b4-4adc-81c2-28402405c9a9","Type":"ContainerDied","Data":"c3474eb4094bf0d37cb4dbeaa8df4894512eaba7946fc556b27c464902273933"} Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.439313 4857 scope.go:117] "RemoveContainer" containerID="012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.440228 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6ffhh" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.477560 4857 scope.go:117] "RemoveContainer" containerID="9875651beb0fb03342e8312ca9a686d6bbdaac46043f90c747db546c18f27426" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.491232 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6ffhh"] Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.507367 4857 scope.go:117] "RemoveContainer" containerID="cd60d9229247d7dd66dc7687b4a18772c0b22a666f1d15d9e49246a35afa4397" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.508885 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6ffhh"] Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.562694 4857 scope.go:117] "RemoveContainer" containerID="012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1" Dec 01 22:19:18 crc kubenswrapper[4857]: E1201 22:19:18.563231 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1\": container with ID starting with 012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1 not found: ID does not exist" containerID="012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.563300 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1"} err="failed to get container status \"012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1\": rpc error: code = NotFound desc = could not find container \"012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1\": container with ID starting with 012eb60064a92bf520b8eb8460fb67a694da360cc5ab8f2da945ec86e55ddee1 not found: ID does not exist" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.563354 4857 scope.go:117] "RemoveContainer" containerID="9875651beb0fb03342e8312ca9a686d6bbdaac46043f90c747db546c18f27426" Dec 01 22:19:18 crc kubenswrapper[4857]: E1201 22:19:18.565496 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9875651beb0fb03342e8312ca9a686d6bbdaac46043f90c747db546c18f27426\": container with ID starting with 9875651beb0fb03342e8312ca9a686d6bbdaac46043f90c747db546c18f27426 not found: ID does not exist" containerID="9875651beb0fb03342e8312ca9a686d6bbdaac46043f90c747db546c18f27426" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.565554 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9875651beb0fb03342e8312ca9a686d6bbdaac46043f90c747db546c18f27426"} err="failed to get container status \"9875651beb0fb03342e8312ca9a686d6bbdaac46043f90c747db546c18f27426\": rpc error: code = NotFound desc = could not find container \"9875651beb0fb03342e8312ca9a686d6bbdaac46043f90c747db546c18f27426\": container with ID starting with 9875651beb0fb03342e8312ca9a686d6bbdaac46043f90c747db546c18f27426 not found: ID does not exist" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.565588 4857 scope.go:117] "RemoveContainer" containerID="cd60d9229247d7dd66dc7687b4a18772c0b22a666f1d15d9e49246a35afa4397" Dec 01 22:19:18 crc kubenswrapper[4857]: E1201 22:19:18.565908 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd60d9229247d7dd66dc7687b4a18772c0b22a666f1d15d9e49246a35afa4397\": container with ID starting with cd60d9229247d7dd66dc7687b4a18772c0b22a666f1d15d9e49246a35afa4397 not found: ID does not exist" containerID="cd60d9229247d7dd66dc7687b4a18772c0b22a666f1d15d9e49246a35afa4397" Dec 01 22:19:18 crc kubenswrapper[4857]: I1201 22:19:18.565933 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd60d9229247d7dd66dc7687b4a18772c0b22a666f1d15d9e49246a35afa4397"} err="failed to get container status \"cd60d9229247d7dd66dc7687b4a18772c0b22a666f1d15d9e49246a35afa4397\": rpc error: code = NotFound desc = could not find container \"cd60d9229247d7dd66dc7687b4a18772c0b22a666f1d15d9e49246a35afa4397\": container with ID starting with cd60d9229247d7dd66dc7687b4a18772c0b22a666f1d15d9e49246a35afa4397 not found: ID does not exist" Dec 01 22:19:19 crc kubenswrapper[4857]: I1201 22:19:19.836471 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:19:19 crc kubenswrapper[4857]: E1201 22:19:19.837118 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:19:19 crc kubenswrapper[4857]: I1201 22:19:19.851678 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcd98f11-32b4-4adc-81c2-28402405c9a9" path="/var/lib/kubelet/pods/dcd98f11-32b4-4adc-81c2-28402405c9a9/volumes" Dec 01 22:19:32 crc kubenswrapper[4857]: I1201 22:19:32.835928 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:19:32 crc kubenswrapper[4857]: E1201 22:19:32.837310 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:19:46 crc kubenswrapper[4857]: I1201 22:19:46.835752 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:19:46 crc kubenswrapper[4857]: E1201 22:19:46.836690 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:19:58 crc kubenswrapper[4857]: I1201 22:19:58.835056 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:19:59 crc kubenswrapper[4857]: I1201 22:19:59.917288 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"1d5a42554e05018f30c89bf3d62c4ed21f317e211a23dbb2f6169d93ade9fe66"} Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.672194 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-txmxn"] Dec 01 22:20:10 crc kubenswrapper[4857]: E1201 22:20:10.675187 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcd98f11-32b4-4adc-81c2-28402405c9a9" containerName="extract-utilities" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.675238 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcd98f11-32b4-4adc-81c2-28402405c9a9" containerName="extract-utilities" Dec 01 22:20:10 crc kubenswrapper[4857]: E1201 22:20:10.675322 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcd98f11-32b4-4adc-81c2-28402405c9a9" containerName="extract-content" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.675334 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcd98f11-32b4-4adc-81c2-28402405c9a9" containerName="extract-content" Dec 01 22:20:10 crc kubenswrapper[4857]: E1201 22:20:10.675419 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcd98f11-32b4-4adc-81c2-28402405c9a9" containerName="registry-server" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.675435 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcd98f11-32b4-4adc-81c2-28402405c9a9" containerName="registry-server" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.678555 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcd98f11-32b4-4adc-81c2-28402405c9a9" containerName="registry-server" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.688822 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.700870 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-txmxn"] Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.804548 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43b50eb1-0058-4ae1-94ec-859b37b746c4-catalog-content\") pod \"redhat-marketplace-txmxn\" (UID: \"43b50eb1-0058-4ae1-94ec-859b37b746c4\") " pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.804875 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43b50eb1-0058-4ae1-94ec-859b37b746c4-utilities\") pod \"redhat-marketplace-txmxn\" (UID: \"43b50eb1-0058-4ae1-94ec-859b37b746c4\") " pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.805013 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdmhw\" (UniqueName: \"kubernetes.io/projected/43b50eb1-0058-4ae1-94ec-859b37b746c4-kube-api-access-zdmhw\") pod \"redhat-marketplace-txmxn\" (UID: \"43b50eb1-0058-4ae1-94ec-859b37b746c4\") " pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.907139 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43b50eb1-0058-4ae1-94ec-859b37b746c4-catalog-content\") pod \"redhat-marketplace-txmxn\" (UID: \"43b50eb1-0058-4ae1-94ec-859b37b746c4\") " pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.907492 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43b50eb1-0058-4ae1-94ec-859b37b746c4-utilities\") pod \"redhat-marketplace-txmxn\" (UID: \"43b50eb1-0058-4ae1-94ec-859b37b746c4\") " pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.907628 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdmhw\" (UniqueName: \"kubernetes.io/projected/43b50eb1-0058-4ae1-94ec-859b37b746c4-kube-api-access-zdmhw\") pod \"redhat-marketplace-txmxn\" (UID: \"43b50eb1-0058-4ae1-94ec-859b37b746c4\") " pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.908405 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43b50eb1-0058-4ae1-94ec-859b37b746c4-catalog-content\") pod \"redhat-marketplace-txmxn\" (UID: \"43b50eb1-0058-4ae1-94ec-859b37b746c4\") " pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.908700 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43b50eb1-0058-4ae1-94ec-859b37b746c4-utilities\") pod \"redhat-marketplace-txmxn\" (UID: \"43b50eb1-0058-4ae1-94ec-859b37b746c4\") " pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:10 crc kubenswrapper[4857]: I1201 22:20:10.924743 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdmhw\" (UniqueName: \"kubernetes.io/projected/43b50eb1-0058-4ae1-94ec-859b37b746c4-kube-api-access-zdmhw\") pod \"redhat-marketplace-txmxn\" (UID: \"43b50eb1-0058-4ae1-94ec-859b37b746c4\") " pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:11 crc kubenswrapper[4857]: I1201 22:20:11.020619 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:11 crc kubenswrapper[4857]: I1201 22:20:11.471902 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-txmxn"] Dec 01 22:20:11 crc kubenswrapper[4857]: W1201 22:20:11.480931 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43b50eb1_0058_4ae1_94ec_859b37b746c4.slice/crio-c8cc477cda5a06fadbd659d3494f3990708f78dbd0c00d83bf1d75c50649132d WatchSource:0}: Error finding container c8cc477cda5a06fadbd659d3494f3990708f78dbd0c00d83bf1d75c50649132d: Status 404 returned error can't find the container with id c8cc477cda5a06fadbd659d3494f3990708f78dbd0c00d83bf1d75c50649132d Dec 01 22:20:12 crc kubenswrapper[4857]: I1201 22:20:12.073775 4857 generic.go:334] "Generic (PLEG): container finished" podID="43b50eb1-0058-4ae1-94ec-859b37b746c4" containerID="6af28353f0612f125433233728a3683369eb92e7ea5da36afd6df9ca4f80c79d" exitCode=0 Dec 01 22:20:12 crc kubenswrapper[4857]: I1201 22:20:12.073996 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txmxn" event={"ID":"43b50eb1-0058-4ae1-94ec-859b37b746c4","Type":"ContainerDied","Data":"6af28353f0612f125433233728a3683369eb92e7ea5da36afd6df9ca4f80c79d"} Dec 01 22:20:12 crc kubenswrapper[4857]: I1201 22:20:12.074301 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txmxn" event={"ID":"43b50eb1-0058-4ae1-94ec-859b37b746c4","Type":"ContainerStarted","Data":"c8cc477cda5a06fadbd659d3494f3990708f78dbd0c00d83bf1d75c50649132d"} Dec 01 22:20:13 crc kubenswrapper[4857]: I1201 22:20:13.086881 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txmxn" event={"ID":"43b50eb1-0058-4ae1-94ec-859b37b746c4","Type":"ContainerStarted","Data":"3d1ce6bef28d28a9281cc4703874933e6b89aab89d619d1b392bb0b5b2416a72"} Dec 01 22:20:14 crc kubenswrapper[4857]: I1201 22:20:14.098833 4857 generic.go:334] "Generic (PLEG): container finished" podID="43b50eb1-0058-4ae1-94ec-859b37b746c4" containerID="3d1ce6bef28d28a9281cc4703874933e6b89aab89d619d1b392bb0b5b2416a72" exitCode=0 Dec 01 22:20:14 crc kubenswrapper[4857]: I1201 22:20:14.098874 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txmxn" event={"ID":"43b50eb1-0058-4ae1-94ec-859b37b746c4","Type":"ContainerDied","Data":"3d1ce6bef28d28a9281cc4703874933e6b89aab89d619d1b392bb0b5b2416a72"} Dec 01 22:20:16 crc kubenswrapper[4857]: I1201 22:20:16.126252 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txmxn" event={"ID":"43b50eb1-0058-4ae1-94ec-859b37b746c4","Type":"ContainerStarted","Data":"2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147"} Dec 01 22:20:16 crc kubenswrapper[4857]: I1201 22:20:16.149377 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-txmxn" podStartSLOduration=3.272120928 podStartE2EDuration="6.149356138s" podCreationTimestamp="2025-12-01 22:20:10 +0000 UTC" firstStartedPulling="2025-12-01 22:20:12.076979615 +0000 UTC m=+2730.567041962" lastFinishedPulling="2025-12-01 22:20:14.954214855 +0000 UTC m=+2733.444277172" observedRunningTime="2025-12-01 22:20:16.147947444 +0000 UTC m=+2734.638009801" watchObservedRunningTime="2025-12-01 22:20:16.149356138 +0000 UTC m=+2734.639418465" Dec 01 22:20:21 crc kubenswrapper[4857]: I1201 22:20:21.020769 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:21 crc kubenswrapper[4857]: I1201 22:20:21.021322 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:21 crc kubenswrapper[4857]: I1201 22:20:21.087955 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:21 crc kubenswrapper[4857]: I1201 22:20:21.290413 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:21 crc kubenswrapper[4857]: I1201 22:20:21.332844 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-txmxn"] Dec 01 22:20:23 crc kubenswrapper[4857]: I1201 22:20:23.266413 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-txmxn" podUID="43b50eb1-0058-4ae1-94ec-859b37b746c4" containerName="registry-server" containerID="cri-o://2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147" gracePeriod=2 Dec 01 22:20:23 crc kubenswrapper[4857]: I1201 22:20:23.803756 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:23 crc kubenswrapper[4857]: I1201 22:20:23.917281 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43b50eb1-0058-4ae1-94ec-859b37b746c4-utilities\") pod \"43b50eb1-0058-4ae1-94ec-859b37b746c4\" (UID: \"43b50eb1-0058-4ae1-94ec-859b37b746c4\") " Dec 01 22:20:23 crc kubenswrapper[4857]: I1201 22:20:23.917342 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43b50eb1-0058-4ae1-94ec-859b37b746c4-catalog-content\") pod \"43b50eb1-0058-4ae1-94ec-859b37b746c4\" (UID: \"43b50eb1-0058-4ae1-94ec-859b37b746c4\") " Dec 01 22:20:23 crc kubenswrapper[4857]: I1201 22:20:23.917535 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdmhw\" (UniqueName: \"kubernetes.io/projected/43b50eb1-0058-4ae1-94ec-859b37b746c4-kube-api-access-zdmhw\") pod \"43b50eb1-0058-4ae1-94ec-859b37b746c4\" (UID: \"43b50eb1-0058-4ae1-94ec-859b37b746c4\") " Dec 01 22:20:23 crc kubenswrapper[4857]: I1201 22:20:23.919174 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43b50eb1-0058-4ae1-94ec-859b37b746c4-utilities" (OuterVolumeSpecName: "utilities") pod "43b50eb1-0058-4ae1-94ec-859b37b746c4" (UID: "43b50eb1-0058-4ae1-94ec-859b37b746c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:20:23 crc kubenswrapper[4857]: I1201 22:20:23.920881 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43b50eb1-0058-4ae1-94ec-859b37b746c4-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:20:23 crc kubenswrapper[4857]: I1201 22:20:23.925208 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43b50eb1-0058-4ae1-94ec-859b37b746c4-kube-api-access-zdmhw" (OuterVolumeSpecName: "kube-api-access-zdmhw") pod "43b50eb1-0058-4ae1-94ec-859b37b746c4" (UID: "43b50eb1-0058-4ae1-94ec-859b37b746c4"). InnerVolumeSpecName "kube-api-access-zdmhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:20:23 crc kubenswrapper[4857]: I1201 22:20:23.945638 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43b50eb1-0058-4ae1-94ec-859b37b746c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43b50eb1-0058-4ae1-94ec-859b37b746c4" (UID: "43b50eb1-0058-4ae1-94ec-859b37b746c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.022509 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43b50eb1-0058-4ae1-94ec-859b37b746c4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.022813 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdmhw\" (UniqueName: \"kubernetes.io/projected/43b50eb1-0058-4ae1-94ec-859b37b746c4-kube-api-access-zdmhw\") on node \"crc\" DevicePath \"\"" Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.282627 4857 generic.go:334] "Generic (PLEG): container finished" podID="43b50eb1-0058-4ae1-94ec-859b37b746c4" containerID="2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147" exitCode=0 Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.282697 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txmxn" event={"ID":"43b50eb1-0058-4ae1-94ec-859b37b746c4","Type":"ContainerDied","Data":"2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147"} Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.282754 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txmxn" Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.282785 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txmxn" event={"ID":"43b50eb1-0058-4ae1-94ec-859b37b746c4","Type":"ContainerDied","Data":"c8cc477cda5a06fadbd659d3494f3990708f78dbd0c00d83bf1d75c50649132d"} Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.282821 4857 scope.go:117] "RemoveContainer" containerID="2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147" Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.314826 4857 scope.go:117] "RemoveContainer" containerID="3d1ce6bef28d28a9281cc4703874933e6b89aab89d619d1b392bb0b5b2416a72" Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.404479 4857 scope.go:117] "RemoveContainer" containerID="6af28353f0612f125433233728a3683369eb92e7ea5da36afd6df9ca4f80c79d" Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.417227 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-txmxn"] Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.436184 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-txmxn"] Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.452445 4857 scope.go:117] "RemoveContainer" containerID="2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147" Dec 01 22:20:24 crc kubenswrapper[4857]: E1201 22:20:24.452941 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147\": container with ID starting with 2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147 not found: ID does not exist" containerID="2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147" Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.453009 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147"} err="failed to get container status \"2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147\": rpc error: code = NotFound desc = could not find container \"2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147\": container with ID starting with 2137dbab1619f19649cc54b0f714b3c9c08a17a4fc39fe7d7f8e773169b29147 not found: ID does not exist" Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.453108 4857 scope.go:117] "RemoveContainer" containerID="3d1ce6bef28d28a9281cc4703874933e6b89aab89d619d1b392bb0b5b2416a72" Dec 01 22:20:24 crc kubenswrapper[4857]: E1201 22:20:24.453652 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d1ce6bef28d28a9281cc4703874933e6b89aab89d619d1b392bb0b5b2416a72\": container with ID starting with 3d1ce6bef28d28a9281cc4703874933e6b89aab89d619d1b392bb0b5b2416a72 not found: ID does not exist" containerID="3d1ce6bef28d28a9281cc4703874933e6b89aab89d619d1b392bb0b5b2416a72" Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.453696 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d1ce6bef28d28a9281cc4703874933e6b89aab89d619d1b392bb0b5b2416a72"} err="failed to get container status \"3d1ce6bef28d28a9281cc4703874933e6b89aab89d619d1b392bb0b5b2416a72\": rpc error: code = NotFound desc = could not find container \"3d1ce6bef28d28a9281cc4703874933e6b89aab89d619d1b392bb0b5b2416a72\": container with ID starting with 3d1ce6bef28d28a9281cc4703874933e6b89aab89d619d1b392bb0b5b2416a72 not found: ID does not exist" Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.453732 4857 scope.go:117] "RemoveContainer" containerID="6af28353f0612f125433233728a3683369eb92e7ea5da36afd6df9ca4f80c79d" Dec 01 22:20:24 crc kubenswrapper[4857]: E1201 22:20:24.454058 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6af28353f0612f125433233728a3683369eb92e7ea5da36afd6df9ca4f80c79d\": container with ID starting with 6af28353f0612f125433233728a3683369eb92e7ea5da36afd6df9ca4f80c79d not found: ID does not exist" containerID="6af28353f0612f125433233728a3683369eb92e7ea5da36afd6df9ca4f80c79d" Dec 01 22:20:24 crc kubenswrapper[4857]: I1201 22:20:24.454098 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6af28353f0612f125433233728a3683369eb92e7ea5da36afd6df9ca4f80c79d"} err="failed to get container status \"6af28353f0612f125433233728a3683369eb92e7ea5da36afd6df9ca4f80c79d\": rpc error: code = NotFound desc = could not find container \"6af28353f0612f125433233728a3683369eb92e7ea5da36afd6df9ca4f80c79d\": container with ID starting with 6af28353f0612f125433233728a3683369eb92e7ea5da36afd6df9ca4f80c79d not found: ID does not exist" Dec 01 22:20:25 crc kubenswrapper[4857]: I1201 22:20:25.848697 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43b50eb1-0058-4ae1-94ec-859b37b746c4" path="/var/lib/kubelet/pods/43b50eb1-0058-4ae1-94ec-859b37b746c4/volumes" Dec 01 22:21:06 crc kubenswrapper[4857]: I1201 22:21:06.908786 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nglww"] Dec 01 22:21:06 crc kubenswrapper[4857]: E1201 22:21:06.910207 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43b50eb1-0058-4ae1-94ec-859b37b746c4" containerName="extract-content" Dec 01 22:21:06 crc kubenswrapper[4857]: I1201 22:21:06.910232 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="43b50eb1-0058-4ae1-94ec-859b37b746c4" containerName="extract-content" Dec 01 22:21:06 crc kubenswrapper[4857]: E1201 22:21:06.910254 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43b50eb1-0058-4ae1-94ec-859b37b746c4" containerName="registry-server" Dec 01 22:21:06 crc kubenswrapper[4857]: I1201 22:21:06.910262 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="43b50eb1-0058-4ae1-94ec-859b37b746c4" containerName="registry-server" Dec 01 22:21:06 crc kubenswrapper[4857]: E1201 22:21:06.910281 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43b50eb1-0058-4ae1-94ec-859b37b746c4" containerName="extract-utilities" Dec 01 22:21:06 crc kubenswrapper[4857]: I1201 22:21:06.910289 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="43b50eb1-0058-4ae1-94ec-859b37b746c4" containerName="extract-utilities" Dec 01 22:21:06 crc kubenswrapper[4857]: I1201 22:21:06.910634 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="43b50eb1-0058-4ae1-94ec-859b37b746c4" containerName="registry-server" Dec 01 22:21:06 crc kubenswrapper[4857]: I1201 22:21:06.912562 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:06 crc kubenswrapper[4857]: I1201 22:21:06.921140 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nglww"] Dec 01 22:21:07 crc kubenswrapper[4857]: I1201 22:21:07.024079 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/425225d2-5fe2-4f15-a295-0241f3b92eef-catalog-content\") pod \"community-operators-nglww\" (UID: \"425225d2-5fe2-4f15-a295-0241f3b92eef\") " pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:07 crc kubenswrapper[4857]: I1201 22:21:07.024468 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65b6c\" (UniqueName: \"kubernetes.io/projected/425225d2-5fe2-4f15-a295-0241f3b92eef-kube-api-access-65b6c\") pod \"community-operators-nglww\" (UID: \"425225d2-5fe2-4f15-a295-0241f3b92eef\") " pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:07 crc kubenswrapper[4857]: I1201 22:21:07.024512 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/425225d2-5fe2-4f15-a295-0241f3b92eef-utilities\") pod \"community-operators-nglww\" (UID: \"425225d2-5fe2-4f15-a295-0241f3b92eef\") " pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:07 crc kubenswrapper[4857]: I1201 22:21:07.125741 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/425225d2-5fe2-4f15-a295-0241f3b92eef-utilities\") pod \"community-operators-nglww\" (UID: \"425225d2-5fe2-4f15-a295-0241f3b92eef\") " pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:07 crc kubenswrapper[4857]: I1201 22:21:07.125887 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/425225d2-5fe2-4f15-a295-0241f3b92eef-catalog-content\") pod \"community-operators-nglww\" (UID: \"425225d2-5fe2-4f15-a295-0241f3b92eef\") " pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:07 crc kubenswrapper[4857]: I1201 22:21:07.125962 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65b6c\" (UniqueName: \"kubernetes.io/projected/425225d2-5fe2-4f15-a295-0241f3b92eef-kube-api-access-65b6c\") pod \"community-operators-nglww\" (UID: \"425225d2-5fe2-4f15-a295-0241f3b92eef\") " pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:07 crc kubenswrapper[4857]: I1201 22:21:07.126341 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/425225d2-5fe2-4f15-a295-0241f3b92eef-utilities\") pod \"community-operators-nglww\" (UID: \"425225d2-5fe2-4f15-a295-0241f3b92eef\") " pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:07 crc kubenswrapper[4857]: I1201 22:21:07.126443 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/425225d2-5fe2-4f15-a295-0241f3b92eef-catalog-content\") pod \"community-operators-nglww\" (UID: \"425225d2-5fe2-4f15-a295-0241f3b92eef\") " pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:07 crc kubenswrapper[4857]: I1201 22:21:07.151192 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65b6c\" (UniqueName: \"kubernetes.io/projected/425225d2-5fe2-4f15-a295-0241f3b92eef-kube-api-access-65b6c\") pod \"community-operators-nglww\" (UID: \"425225d2-5fe2-4f15-a295-0241f3b92eef\") " pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:07 crc kubenswrapper[4857]: I1201 22:21:07.253660 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:07 crc kubenswrapper[4857]: I1201 22:21:07.791971 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nglww"] Dec 01 22:21:07 crc kubenswrapper[4857]: I1201 22:21:07.828385 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nglww" event={"ID":"425225d2-5fe2-4f15-a295-0241f3b92eef","Type":"ContainerStarted","Data":"9337e520b7fa06a14f9b5d6bb39b638dc50daf844dd5e05098b1b8ada377e593"} Dec 01 22:21:08 crc kubenswrapper[4857]: I1201 22:21:08.841453 4857 generic.go:334] "Generic (PLEG): container finished" podID="425225d2-5fe2-4f15-a295-0241f3b92eef" containerID="cc5dba095a1397da1e59b00df37ea348ad960452cc3b6d27447da8db834caaf9" exitCode=0 Dec 01 22:21:08 crc kubenswrapper[4857]: I1201 22:21:08.841676 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nglww" event={"ID":"425225d2-5fe2-4f15-a295-0241f3b92eef","Type":"ContainerDied","Data":"cc5dba095a1397da1e59b00df37ea348ad960452cc3b6d27447da8db834caaf9"} Dec 01 22:21:09 crc kubenswrapper[4857]: I1201 22:21:09.853019 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nglww" event={"ID":"425225d2-5fe2-4f15-a295-0241f3b92eef","Type":"ContainerStarted","Data":"39760b8d6367926b056ae0f6aba58f013f3a7e6b66c747c79dc5c9f63d8673ef"} Dec 01 22:21:10 crc kubenswrapper[4857]: I1201 22:21:10.870360 4857 generic.go:334] "Generic (PLEG): container finished" podID="425225d2-5fe2-4f15-a295-0241f3b92eef" containerID="39760b8d6367926b056ae0f6aba58f013f3a7e6b66c747c79dc5c9f63d8673ef" exitCode=0 Dec 01 22:21:10 crc kubenswrapper[4857]: I1201 22:21:10.870431 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nglww" event={"ID":"425225d2-5fe2-4f15-a295-0241f3b92eef","Type":"ContainerDied","Data":"39760b8d6367926b056ae0f6aba58f013f3a7e6b66c747c79dc5c9f63d8673ef"} Dec 01 22:21:11 crc kubenswrapper[4857]: I1201 22:21:11.888096 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nglww" event={"ID":"425225d2-5fe2-4f15-a295-0241f3b92eef","Type":"ContainerStarted","Data":"b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1"} Dec 01 22:21:11 crc kubenswrapper[4857]: I1201 22:21:11.919014 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nglww" podStartSLOduration=3.438245993 podStartE2EDuration="5.918988057s" podCreationTimestamp="2025-12-01 22:21:06 +0000 UTC" firstStartedPulling="2025-12-01 22:21:08.844537743 +0000 UTC m=+2787.334600100" lastFinishedPulling="2025-12-01 22:21:11.325279807 +0000 UTC m=+2789.815342164" observedRunningTime="2025-12-01 22:21:11.914855335 +0000 UTC m=+2790.404917662" watchObservedRunningTime="2025-12-01 22:21:11.918988057 +0000 UTC m=+2790.409050414" Dec 01 22:21:15 crc kubenswrapper[4857]: I1201 22:21:15.936384 4857 generic.go:334] "Generic (PLEG): container finished" podID="3b8d39f4-2804-4fda-bec9-0cd4c2401f42" containerID="d1622483f2911e92330e1f5a81f0bcf6e957c9c14981300099800e8d58a2467a" exitCode=0 Dec 01 22:21:15 crc kubenswrapper[4857]: I1201 22:21:15.936475 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" event={"ID":"3b8d39f4-2804-4fda-bec9-0cd4c2401f42","Type":"ContainerDied","Data":"d1622483f2911e92330e1f5a81f0bcf6e957c9c14981300099800e8d58a2467a"} Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.254269 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.254717 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.352733 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.456638 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.669504 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-inventory\") pod \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.669570 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-cell1-compute-config-1\") pod \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.669631 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krxb4\" (UniqueName: \"kubernetes.io/projected/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-kube-api-access-krxb4\") pod \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.669728 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-ssh-key\") pod \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.669761 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-migration-ssh-key-0\") pod \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.669808 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-combined-ca-bundle\") pod \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.669933 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-migration-ssh-key-1\") pod \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.669950 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-cell1-compute-config-0\") pod \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.669992 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-extra-config-0\") pod \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\" (UID: \"3b8d39f4-2804-4fda-bec9-0cd4c2401f42\") " Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.689581 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "3b8d39f4-2804-4fda-bec9-0cd4c2401f42" (UID: "3b8d39f4-2804-4fda-bec9-0cd4c2401f42"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.691644 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-kube-api-access-krxb4" (OuterVolumeSpecName: "kube-api-access-krxb4") pod "3b8d39f4-2804-4fda-bec9-0cd4c2401f42" (UID: "3b8d39f4-2804-4fda-bec9-0cd4c2401f42"). InnerVolumeSpecName "kube-api-access-krxb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.712659 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "3b8d39f4-2804-4fda-bec9-0cd4c2401f42" (UID: "3b8d39f4-2804-4fda-bec9-0cd4c2401f42"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.716360 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "3b8d39f4-2804-4fda-bec9-0cd4c2401f42" (UID: "3b8d39f4-2804-4fda-bec9-0cd4c2401f42"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.728760 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-inventory" (OuterVolumeSpecName: "inventory") pod "3b8d39f4-2804-4fda-bec9-0cd4c2401f42" (UID: "3b8d39f4-2804-4fda-bec9-0cd4c2401f42"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.728896 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "3b8d39f4-2804-4fda-bec9-0cd4c2401f42" (UID: "3b8d39f4-2804-4fda-bec9-0cd4c2401f42"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.733558 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "3b8d39f4-2804-4fda-bec9-0cd4c2401f42" (UID: "3b8d39f4-2804-4fda-bec9-0cd4c2401f42"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.733816 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "3b8d39f4-2804-4fda-bec9-0cd4c2401f42" (UID: "3b8d39f4-2804-4fda-bec9-0cd4c2401f42"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.750248 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3b8d39f4-2804-4fda-bec9-0cd4c2401f42" (UID: "3b8d39f4-2804-4fda-bec9-0cd4c2401f42"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.772825 4857 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.772860 4857 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.772871 4857 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.772884 4857 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.772893 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.772902 4857 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.772910 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krxb4\" (UniqueName: \"kubernetes.io/projected/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-kube-api-access-krxb4\") on node \"crc\" DevicePath \"\"" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.772918 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.772927 4857 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3b8d39f4-2804-4fda-bec9-0cd4c2401f42-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.957372 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" event={"ID":"3b8d39f4-2804-4fda-bec9-0cd4c2401f42","Type":"ContainerDied","Data":"db5f9059414aae02ff33e6b32db44d0ae706c6ceadfb358182cd4af16a5f7cde"} Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.957422 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db5f9059414aae02ff33e6b32db44d0ae706c6ceadfb358182cd4af16a5f7cde" Dec 01 22:21:17 crc kubenswrapper[4857]: I1201 22:21:17.957427 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l7ltc" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.060764 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.077547 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms"] Dec 01 22:21:18 crc kubenswrapper[4857]: E1201 22:21:18.078081 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8d39f4-2804-4fda-bec9-0cd4c2401f42" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.078104 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8d39f4-2804-4fda-bec9-0cd4c2401f42" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.078423 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8d39f4-2804-4fda-bec9-0cd4c2401f42" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.079259 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.083693 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.083743 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.083876 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.083941 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9kbl" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.090175 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.108503 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms"] Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.164999 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nglww"] Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.282911 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.283310 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.283384 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.283487 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4lv8\" (UniqueName: \"kubernetes.io/projected/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-kube-api-access-b4lv8\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.283541 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.283568 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.283662 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.385413 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.385712 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.385836 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.385935 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4lv8\" (UniqueName: \"kubernetes.io/projected/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-kube-api-access-b4lv8\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.386085 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.386464 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.386655 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.395849 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.395866 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.396302 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.396579 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.396845 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.400733 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.410823 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4lv8\" (UniqueName: \"kubernetes.io/projected/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-kube-api-access-b4lv8\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6wnms\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:18 crc kubenswrapper[4857]: I1201 22:21:18.697557 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:21:19 crc kubenswrapper[4857]: W1201 22:21:19.318066 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2aa8b0d6_1236_48c6_a69c_5996c4a0d6d6.slice/crio-a14bb4f81d35bb93827e5a936db68ac3be212f19cf24aba3585d509e4c0415e9 WatchSource:0}: Error finding container a14bb4f81d35bb93827e5a936db68ac3be212f19cf24aba3585d509e4c0415e9: Status 404 returned error can't find the container with id a14bb4f81d35bb93827e5a936db68ac3be212f19cf24aba3585d509e4c0415e9 Dec 01 22:21:19 crc kubenswrapper[4857]: I1201 22:21:19.320205 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms"] Dec 01 22:21:19 crc kubenswrapper[4857]: I1201 22:21:19.980125 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" event={"ID":"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6","Type":"ContainerStarted","Data":"a14bb4f81d35bb93827e5a936db68ac3be212f19cf24aba3585d509e4c0415e9"} Dec 01 22:21:19 crc kubenswrapper[4857]: I1201 22:21:19.980451 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nglww" podUID="425225d2-5fe2-4f15-a295-0241f3b92eef" containerName="registry-server" containerID="cri-o://b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1" gracePeriod=2 Dec 01 22:21:20 crc kubenswrapper[4857]: I1201 22:21:20.576725 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:20 crc kubenswrapper[4857]: I1201 22:21:20.738737 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65b6c\" (UniqueName: \"kubernetes.io/projected/425225d2-5fe2-4f15-a295-0241f3b92eef-kube-api-access-65b6c\") pod \"425225d2-5fe2-4f15-a295-0241f3b92eef\" (UID: \"425225d2-5fe2-4f15-a295-0241f3b92eef\") " Dec 01 22:21:20 crc kubenswrapper[4857]: I1201 22:21:20.739273 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/425225d2-5fe2-4f15-a295-0241f3b92eef-utilities\") pod \"425225d2-5fe2-4f15-a295-0241f3b92eef\" (UID: \"425225d2-5fe2-4f15-a295-0241f3b92eef\") " Dec 01 22:21:20 crc kubenswrapper[4857]: I1201 22:21:20.739316 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/425225d2-5fe2-4f15-a295-0241f3b92eef-catalog-content\") pod \"425225d2-5fe2-4f15-a295-0241f3b92eef\" (UID: \"425225d2-5fe2-4f15-a295-0241f3b92eef\") " Dec 01 22:21:20 crc kubenswrapper[4857]: I1201 22:21:20.740864 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/425225d2-5fe2-4f15-a295-0241f3b92eef-utilities" (OuterVolumeSpecName: "utilities") pod "425225d2-5fe2-4f15-a295-0241f3b92eef" (UID: "425225d2-5fe2-4f15-a295-0241f3b92eef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:21:20 crc kubenswrapper[4857]: I1201 22:21:20.746196 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/425225d2-5fe2-4f15-a295-0241f3b92eef-kube-api-access-65b6c" (OuterVolumeSpecName: "kube-api-access-65b6c") pod "425225d2-5fe2-4f15-a295-0241f3b92eef" (UID: "425225d2-5fe2-4f15-a295-0241f3b92eef"). InnerVolumeSpecName "kube-api-access-65b6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:21:20 crc kubenswrapper[4857]: I1201 22:21:20.798469 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/425225d2-5fe2-4f15-a295-0241f3b92eef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "425225d2-5fe2-4f15-a295-0241f3b92eef" (UID: "425225d2-5fe2-4f15-a295-0241f3b92eef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:21:20 crc kubenswrapper[4857]: I1201 22:21:20.842255 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65b6c\" (UniqueName: \"kubernetes.io/projected/425225d2-5fe2-4f15-a295-0241f3b92eef-kube-api-access-65b6c\") on node \"crc\" DevicePath \"\"" Dec 01 22:21:20 crc kubenswrapper[4857]: I1201 22:21:20.842282 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/425225d2-5fe2-4f15-a295-0241f3b92eef-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:21:20 crc kubenswrapper[4857]: I1201 22:21:20.842293 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/425225d2-5fe2-4f15-a295-0241f3b92eef-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:21:20 crc kubenswrapper[4857]: I1201 22:21:20.997829 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" event={"ID":"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6","Type":"ContainerStarted","Data":"7ceac6562434537cfb6750c61ade2ab1365d4e3ebde79c9766789030ad70d19b"} Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.006664 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nglww" Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.006673 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nglww" event={"ID":"425225d2-5fe2-4f15-a295-0241f3b92eef","Type":"ContainerDied","Data":"b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1"} Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.006754 4857 scope.go:117] "RemoveContainer" containerID="b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1" Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.006494 4857 generic.go:334] "Generic (PLEG): container finished" podID="425225d2-5fe2-4f15-a295-0241f3b92eef" containerID="b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1" exitCode=0 Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.012653 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nglww" event={"ID":"425225d2-5fe2-4f15-a295-0241f3b92eef","Type":"ContainerDied","Data":"9337e520b7fa06a14f9b5d6bb39b638dc50daf844dd5e05098b1b8ada377e593"} Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.039375 4857 scope.go:117] "RemoveContainer" containerID="39760b8d6367926b056ae0f6aba58f013f3a7e6b66c747c79dc5c9f63d8673ef" Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.039587 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" podStartSLOduration=2.214445546 podStartE2EDuration="3.039564443s" podCreationTimestamp="2025-12-01 22:21:18 +0000 UTC" firstStartedPulling="2025-12-01 22:21:19.321680301 +0000 UTC m=+2797.811742618" lastFinishedPulling="2025-12-01 22:21:20.146799188 +0000 UTC m=+2798.636861515" observedRunningTime="2025-12-01 22:21:21.020342549 +0000 UTC m=+2799.510404866" watchObservedRunningTime="2025-12-01 22:21:21.039564443 +0000 UTC m=+2799.529626770" Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.056614 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nglww"] Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.069360 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nglww"] Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.082262 4857 scope.go:117] "RemoveContainer" containerID="cc5dba095a1397da1e59b00df37ea348ad960452cc3b6d27447da8db834caaf9" Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.106359 4857 scope.go:117] "RemoveContainer" containerID="b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1" Dec 01 22:21:21 crc kubenswrapper[4857]: E1201 22:21:21.106871 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1\": container with ID starting with b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1 not found: ID does not exist" containerID="b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1" Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.106939 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1"} err="failed to get container status \"b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1\": rpc error: code = NotFound desc = could not find container \"b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1\": container with ID starting with b56b440f23ee2211a75c9cb92d9d8e189f9e1613c5c815a16283df74c1ad49d1 not found: ID does not exist" Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.106982 4857 scope.go:117] "RemoveContainer" containerID="39760b8d6367926b056ae0f6aba58f013f3a7e6b66c747c79dc5c9f63d8673ef" Dec 01 22:21:21 crc kubenswrapper[4857]: E1201 22:21:21.107394 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39760b8d6367926b056ae0f6aba58f013f3a7e6b66c747c79dc5c9f63d8673ef\": container with ID starting with 39760b8d6367926b056ae0f6aba58f013f3a7e6b66c747c79dc5c9f63d8673ef not found: ID does not exist" containerID="39760b8d6367926b056ae0f6aba58f013f3a7e6b66c747c79dc5c9f63d8673ef" Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.107443 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39760b8d6367926b056ae0f6aba58f013f3a7e6b66c747c79dc5c9f63d8673ef"} err="failed to get container status \"39760b8d6367926b056ae0f6aba58f013f3a7e6b66c747c79dc5c9f63d8673ef\": rpc error: code = NotFound desc = could not find container \"39760b8d6367926b056ae0f6aba58f013f3a7e6b66c747c79dc5c9f63d8673ef\": container with ID starting with 39760b8d6367926b056ae0f6aba58f013f3a7e6b66c747c79dc5c9f63d8673ef not found: ID does not exist" Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.107473 4857 scope.go:117] "RemoveContainer" containerID="cc5dba095a1397da1e59b00df37ea348ad960452cc3b6d27447da8db834caaf9" Dec 01 22:21:21 crc kubenswrapper[4857]: E1201 22:21:21.107947 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc5dba095a1397da1e59b00df37ea348ad960452cc3b6d27447da8db834caaf9\": container with ID starting with cc5dba095a1397da1e59b00df37ea348ad960452cc3b6d27447da8db834caaf9 not found: ID does not exist" containerID="cc5dba095a1397da1e59b00df37ea348ad960452cc3b6d27447da8db834caaf9" Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.108075 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc5dba095a1397da1e59b00df37ea348ad960452cc3b6d27447da8db834caaf9"} err="failed to get container status \"cc5dba095a1397da1e59b00df37ea348ad960452cc3b6d27447da8db834caaf9\": rpc error: code = NotFound desc = could not find container \"cc5dba095a1397da1e59b00df37ea348ad960452cc3b6d27447da8db834caaf9\": container with ID starting with cc5dba095a1397da1e59b00df37ea348ad960452cc3b6d27447da8db834caaf9 not found: ID does not exist" Dec 01 22:21:21 crc kubenswrapper[4857]: I1201 22:21:21.854435 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="425225d2-5fe2-4f15-a295-0241f3b92eef" path="/var/lib/kubelet/pods/425225d2-5fe2-4f15-a295-0241f3b92eef/volumes" Dec 01 22:22:27 crc kubenswrapper[4857]: I1201 22:22:27.806323 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:22:27 crc kubenswrapper[4857]: I1201 22:22:27.806858 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:22:57 crc kubenswrapper[4857]: I1201 22:22:57.806654 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:22:57 crc kubenswrapper[4857]: I1201 22:22:57.807723 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:23:21 crc kubenswrapper[4857]: I1201 22:23:21.821134 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qd6tb"] Dec 01 22:23:21 crc kubenswrapper[4857]: E1201 22:23:21.822399 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="425225d2-5fe2-4f15-a295-0241f3b92eef" containerName="extract-utilities" Dec 01 22:23:21 crc kubenswrapper[4857]: I1201 22:23:21.822422 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="425225d2-5fe2-4f15-a295-0241f3b92eef" containerName="extract-utilities" Dec 01 22:23:21 crc kubenswrapper[4857]: E1201 22:23:21.822462 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="425225d2-5fe2-4f15-a295-0241f3b92eef" containerName="registry-server" Dec 01 22:23:21 crc kubenswrapper[4857]: I1201 22:23:21.822474 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="425225d2-5fe2-4f15-a295-0241f3b92eef" containerName="registry-server" Dec 01 22:23:21 crc kubenswrapper[4857]: E1201 22:23:21.822529 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="425225d2-5fe2-4f15-a295-0241f3b92eef" containerName="extract-content" Dec 01 22:23:21 crc kubenswrapper[4857]: I1201 22:23:21.822544 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="425225d2-5fe2-4f15-a295-0241f3b92eef" containerName="extract-content" Dec 01 22:23:21 crc kubenswrapper[4857]: I1201 22:23:21.822905 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="425225d2-5fe2-4f15-a295-0241f3b92eef" containerName="registry-server" Dec 01 22:23:21 crc kubenswrapper[4857]: I1201 22:23:21.825324 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:21 crc kubenswrapper[4857]: I1201 22:23:21.854991 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qd6tb"] Dec 01 22:23:21 crc kubenswrapper[4857]: I1201 22:23:21.923229 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-catalog-content\") pod \"certified-operators-qd6tb\" (UID: \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\") " pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:21 crc kubenswrapper[4857]: I1201 22:23:21.923278 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8p8n\" (UniqueName: \"kubernetes.io/projected/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-kube-api-access-p8p8n\") pod \"certified-operators-qd6tb\" (UID: \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\") " pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:21 crc kubenswrapper[4857]: I1201 22:23:21.923309 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-utilities\") pod \"certified-operators-qd6tb\" (UID: \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\") " pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:22 crc kubenswrapper[4857]: I1201 22:23:22.025757 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-catalog-content\") pod \"certified-operators-qd6tb\" (UID: \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\") " pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:22 crc kubenswrapper[4857]: I1201 22:23:22.025832 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-catalog-content\") pod \"certified-operators-qd6tb\" (UID: \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\") " pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:22 crc kubenswrapper[4857]: I1201 22:23:22.025883 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8p8n\" (UniqueName: \"kubernetes.io/projected/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-kube-api-access-p8p8n\") pod \"certified-operators-qd6tb\" (UID: \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\") " pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:22 crc kubenswrapper[4857]: I1201 22:23:22.026316 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-utilities\") pod \"certified-operators-qd6tb\" (UID: \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\") " pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:22 crc kubenswrapper[4857]: I1201 22:23:22.026701 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-utilities\") pod \"certified-operators-qd6tb\" (UID: \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\") " pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:22 crc kubenswrapper[4857]: I1201 22:23:22.065938 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8p8n\" (UniqueName: \"kubernetes.io/projected/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-kube-api-access-p8p8n\") pod \"certified-operators-qd6tb\" (UID: \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\") " pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:22 crc kubenswrapper[4857]: I1201 22:23:22.164589 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:22 crc kubenswrapper[4857]: I1201 22:23:22.673991 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qd6tb"] Dec 01 22:23:23 crc kubenswrapper[4857]: I1201 22:23:23.507909 4857 generic.go:334] "Generic (PLEG): container finished" podID="8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" containerID="bcc09cc92b460770758f3386ee47427332090208240c88b4ae69c369744c60ca" exitCode=0 Dec 01 22:23:23 crc kubenswrapper[4857]: I1201 22:23:23.508003 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qd6tb" event={"ID":"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9","Type":"ContainerDied","Data":"bcc09cc92b460770758f3386ee47427332090208240c88b4ae69c369744c60ca"} Dec 01 22:23:23 crc kubenswrapper[4857]: I1201 22:23:23.508364 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qd6tb" event={"ID":"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9","Type":"ContainerStarted","Data":"103e28f4f6bc64ffa5f7038f78c59c292e60aa70d1d7de9dfd7838a161c93eec"} Dec 01 22:23:23 crc kubenswrapper[4857]: I1201 22:23:23.526697 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 22:23:24 crc kubenswrapper[4857]: I1201 22:23:24.523414 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qd6tb" event={"ID":"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9","Type":"ContainerStarted","Data":"99fa09734d6859dea0b7d5628e7c2b1a1d0de6adc00243a6343971e35cba79f8"} Dec 01 22:23:25 crc kubenswrapper[4857]: I1201 22:23:25.537184 4857 generic.go:334] "Generic (PLEG): container finished" podID="8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" containerID="99fa09734d6859dea0b7d5628e7c2b1a1d0de6adc00243a6343971e35cba79f8" exitCode=0 Dec 01 22:23:25 crc kubenswrapper[4857]: I1201 22:23:25.537338 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qd6tb" event={"ID":"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9","Type":"ContainerDied","Data":"99fa09734d6859dea0b7d5628e7c2b1a1d0de6adc00243a6343971e35cba79f8"} Dec 01 22:23:26 crc kubenswrapper[4857]: I1201 22:23:26.551393 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qd6tb" event={"ID":"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9","Type":"ContainerStarted","Data":"722642a2b92ce978766499c63e078bc163e38c7e3cbb6e4e4647ebcf2973ddfb"} Dec 01 22:23:26 crc kubenswrapper[4857]: I1201 22:23:26.607924 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qd6tb" podStartSLOduration=2.8408797789999998 podStartE2EDuration="5.607905662s" podCreationTimestamp="2025-12-01 22:23:21 +0000 UTC" firstStartedPulling="2025-12-01 22:23:23.526385164 +0000 UTC m=+2922.016447491" lastFinishedPulling="2025-12-01 22:23:26.293411047 +0000 UTC m=+2924.783473374" observedRunningTime="2025-12-01 22:23:26.603675827 +0000 UTC m=+2925.093738164" watchObservedRunningTime="2025-12-01 22:23:26.607905662 +0000 UTC m=+2925.097967989" Dec 01 22:23:27 crc kubenswrapper[4857]: I1201 22:23:27.806311 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:23:27 crc kubenswrapper[4857]: I1201 22:23:27.806393 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:23:27 crc kubenswrapper[4857]: I1201 22:23:27.806446 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 22:23:27 crc kubenswrapper[4857]: I1201 22:23:27.807376 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d5a42554e05018f30c89bf3d62c4ed21f317e211a23dbb2f6169d93ade9fe66"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 22:23:27 crc kubenswrapper[4857]: I1201 22:23:27.807454 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://1d5a42554e05018f30c89bf3d62c4ed21f317e211a23dbb2f6169d93ade9fe66" gracePeriod=600 Dec 01 22:23:28 crc kubenswrapper[4857]: I1201 22:23:28.573504 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="1d5a42554e05018f30c89bf3d62c4ed21f317e211a23dbb2f6169d93ade9fe66" exitCode=0 Dec 01 22:23:28 crc kubenswrapper[4857]: I1201 22:23:28.573582 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"1d5a42554e05018f30c89bf3d62c4ed21f317e211a23dbb2f6169d93ade9fe66"} Dec 01 22:23:28 crc kubenswrapper[4857]: I1201 22:23:28.574148 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170"} Dec 01 22:23:28 crc kubenswrapper[4857]: I1201 22:23:28.574183 4857 scope.go:117] "RemoveContainer" containerID="d8c30fa18d2ff421ba9df1a1392df4db3b0a178161d8da4d6e03afe1c36b6159" Dec 01 22:23:32 crc kubenswrapper[4857]: I1201 22:23:32.165037 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:32 crc kubenswrapper[4857]: I1201 22:23:32.165939 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:32 crc kubenswrapper[4857]: I1201 22:23:32.262707 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:32 crc kubenswrapper[4857]: I1201 22:23:32.699022 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:32 crc kubenswrapper[4857]: I1201 22:23:32.786530 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qd6tb"] Dec 01 22:23:34 crc kubenswrapper[4857]: I1201 22:23:34.649699 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qd6tb" podUID="8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" containerName="registry-server" containerID="cri-o://722642a2b92ce978766499c63e078bc163e38c7e3cbb6e4e4647ebcf2973ddfb" gracePeriod=2 Dec 01 22:23:35 crc kubenswrapper[4857]: I1201 22:23:35.669806 4857 generic.go:334] "Generic (PLEG): container finished" podID="8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" containerID="722642a2b92ce978766499c63e078bc163e38c7e3cbb6e4e4647ebcf2973ddfb" exitCode=0 Dec 01 22:23:35 crc kubenswrapper[4857]: I1201 22:23:35.670579 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qd6tb" event={"ID":"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9","Type":"ContainerDied","Data":"722642a2b92ce978766499c63e078bc163e38c7e3cbb6e4e4647ebcf2973ddfb"} Dec 01 22:23:35 crc kubenswrapper[4857]: I1201 22:23:35.670613 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qd6tb" event={"ID":"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9","Type":"ContainerDied","Data":"103e28f4f6bc64ffa5f7038f78c59c292e60aa70d1d7de9dfd7838a161c93eec"} Dec 01 22:23:35 crc kubenswrapper[4857]: I1201 22:23:35.670632 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="103e28f4f6bc64ffa5f7038f78c59c292e60aa70d1d7de9dfd7838a161c93eec" Dec 01 22:23:35 crc kubenswrapper[4857]: I1201 22:23:35.766258 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:35 crc kubenswrapper[4857]: I1201 22:23:35.899862 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-utilities\") pod \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\" (UID: \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\") " Dec 01 22:23:35 crc kubenswrapper[4857]: I1201 22:23:35.899937 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8p8n\" (UniqueName: \"kubernetes.io/projected/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-kube-api-access-p8p8n\") pod \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\" (UID: \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\") " Dec 01 22:23:35 crc kubenswrapper[4857]: I1201 22:23:35.900067 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-catalog-content\") pod \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\" (UID: \"8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9\") " Dec 01 22:23:35 crc kubenswrapper[4857]: I1201 22:23:35.901921 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-utilities" (OuterVolumeSpecName: "utilities") pod "8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" (UID: "8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:23:35 crc kubenswrapper[4857]: I1201 22:23:35.910610 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-kube-api-access-p8p8n" (OuterVolumeSpecName: "kube-api-access-p8p8n") pod "8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" (UID: "8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9"). InnerVolumeSpecName "kube-api-access-p8p8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:23:35 crc kubenswrapper[4857]: I1201 22:23:35.989671 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" (UID: "8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:23:36 crc kubenswrapper[4857]: I1201 22:23:36.002105 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:23:36 crc kubenswrapper[4857]: I1201 22:23:36.002139 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8p8n\" (UniqueName: \"kubernetes.io/projected/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-kube-api-access-p8p8n\") on node \"crc\" DevicePath \"\"" Dec 01 22:23:36 crc kubenswrapper[4857]: I1201 22:23:36.002152 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:23:36 crc kubenswrapper[4857]: I1201 22:23:36.683497 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qd6tb" Dec 01 22:23:36 crc kubenswrapper[4857]: I1201 22:23:36.754775 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qd6tb"] Dec 01 22:23:36 crc kubenswrapper[4857]: I1201 22:23:36.768490 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qd6tb"] Dec 01 22:23:37 crc kubenswrapper[4857]: I1201 22:23:37.908893 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" path="/var/lib/kubelet/pods/8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9/volumes" Dec 01 22:24:10 crc kubenswrapper[4857]: I1201 22:24:10.062003 4857 generic.go:334] "Generic (PLEG): container finished" podID="2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6" containerID="7ceac6562434537cfb6750c61ade2ab1365d4e3ebde79c9766789030ad70d19b" exitCode=0 Dec 01 22:24:10 crc kubenswrapper[4857]: I1201 22:24:10.062082 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" event={"ID":"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6","Type":"ContainerDied","Data":"7ceac6562434537cfb6750c61ade2ab1365d4e3ebde79c9766789030ad70d19b"} Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.603121 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.615020 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-2\") pod \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.615295 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-inventory\") pod \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.615376 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-0\") pod \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.615449 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4lv8\" (UniqueName: \"kubernetes.io/projected/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-kube-api-access-b4lv8\") pod \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.615503 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-telemetry-combined-ca-bundle\") pod \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.615961 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ssh-key\") pod \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.616084 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-1\") pod \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\" (UID: \"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6\") " Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.621118 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-kube-api-access-b4lv8" (OuterVolumeSpecName: "kube-api-access-b4lv8") pod "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6" (UID: "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6"). InnerVolumeSpecName "kube-api-access-b4lv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.622868 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6" (UID: "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.670746 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-inventory" (OuterVolumeSpecName: "inventory") pod "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6" (UID: "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.670821 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6" (UID: "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.676339 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6" (UID: "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.693183 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6" (UID: "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.710427 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6" (UID: "2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.719136 4857 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.719180 4857 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.719196 4857 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.719211 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4lv8\" (UniqueName: \"kubernetes.io/projected/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-kube-api-access-b4lv8\") on node \"crc\" DevicePath \"\"" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.719225 4857 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.719235 4857 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 22:24:11 crc kubenswrapper[4857]: I1201 22:24:11.719246 4857 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 01 22:24:12 crc kubenswrapper[4857]: I1201 22:24:12.100878 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" event={"ID":"2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6","Type":"ContainerDied","Data":"a14bb4f81d35bb93827e5a936db68ac3be212f19cf24aba3585d509e4c0415e9"} Dec 01 22:24:12 crc kubenswrapper[4857]: I1201 22:24:12.100960 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a14bb4f81d35bb93827e5a936db68ac3be212f19cf24aba3585d509e4c0415e9" Dec 01 22:24:12 crc kubenswrapper[4857]: I1201 22:24:12.100967 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6wnms" Dec 01 22:25:57 crc kubenswrapper[4857]: I1201 22:25:57.806835 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:25:57 crc kubenswrapper[4857]: I1201 22:25:57.807671 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:26:27 crc kubenswrapper[4857]: I1201 22:26:27.806797 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:26:27 crc kubenswrapper[4857]: I1201 22:26:27.807330 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:26:57 crc kubenswrapper[4857]: I1201 22:26:57.805946 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:26:57 crc kubenswrapper[4857]: I1201 22:26:57.806680 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:26:57 crc kubenswrapper[4857]: I1201 22:26:57.806751 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 22:26:57 crc kubenswrapper[4857]: I1201 22:26:57.807591 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 22:26:57 crc kubenswrapper[4857]: I1201 22:26:57.807701 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" gracePeriod=600 Dec 01 22:26:57 crc kubenswrapper[4857]: E1201 22:26:57.934467 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:26:58 crc kubenswrapper[4857]: I1201 22:26:58.173509 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" exitCode=0 Dec 01 22:26:58 crc kubenswrapper[4857]: I1201 22:26:58.173616 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170"} Dec 01 22:26:58 crc kubenswrapper[4857]: I1201 22:26:58.173911 4857 scope.go:117] "RemoveContainer" containerID="1d5a42554e05018f30c89bf3d62c4ed21f317e211a23dbb2f6169d93ade9fe66" Dec 01 22:26:58 crc kubenswrapper[4857]: I1201 22:26:58.174784 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:26:58 crc kubenswrapper[4857]: E1201 22:26:58.175300 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:26:58 crc kubenswrapper[4857]: I1201 22:26:58.710957 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f765fc7d6-x797d_ff230e58-1711-4638-89d6-daf91fb3bfc1/manager/0.log" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.723779 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.724361 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="6821156d-2d04-4608-a5c7-c70b1058fdc0" containerName="openstackclient" containerID="cri-o://c657db47da331af1813f6ade9364da74d4954e848ff4cf4cb00cf13198f57d1d" gracePeriod=2 Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.733060 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.800854 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 22:27:00 crc kubenswrapper[4857]: E1201 22:27:00.801773 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" containerName="extract-utilities" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.801798 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" containerName="extract-utilities" Dec 01 22:27:00 crc kubenswrapper[4857]: E1201 22:27:00.801815 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" containerName="registry-server" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.801824 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" containerName="registry-server" Dec 01 22:27:00 crc kubenswrapper[4857]: E1201 22:27:00.801852 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" containerName="extract-content" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.801861 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" containerName="extract-content" Dec 01 22:27:00 crc kubenswrapper[4857]: E1201 22:27:00.801877 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6821156d-2d04-4608-a5c7-c70b1058fdc0" containerName="openstackclient" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.801885 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="6821156d-2d04-4608-a5c7-c70b1058fdc0" containerName="openstackclient" Dec 01 22:27:00 crc kubenswrapper[4857]: E1201 22:27:00.801926 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.801939 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.802198 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.802219 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="6821156d-2d04-4608-a5c7-c70b1058fdc0" containerName="openstackclient" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.802230 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="8817a6be-e2f4-4a1d-ab3e-df8e1631b7e9" containerName="registry-server" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.803340 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.826576 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.841898 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 22:27:00 crc kubenswrapper[4857]: E1201 22:27:00.842790 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-brxts openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[combined-ca-bundle kube-api-access-brxts openstack-config openstack-config-secret]: context canceled" pod="openstack/openstackclient" podUID="4ea60739-fcd8-45c6-9353-b44dd058f404" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.849482 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.873644 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.875490 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.885074 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.885725 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="6821156d-2d04-4608-a5c7-c70b1058fdc0" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.893429 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="4ea60739-fcd8-45c6-9353-b44dd058f404" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.996374 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49nzh\" (UniqueName: \"kubernetes.io/projected/2f50f03d-e84a-4c7e-a646-047bcfd3046f-kube-api-access-49nzh\") pod \"openstackclient\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " pod="openstack/openstackclient" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.996480 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2f50f03d-e84a-4c7e-a646-047bcfd3046f-openstack-config-secret\") pod \"openstackclient\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " pod="openstack/openstackclient" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.996527 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2f50f03d-e84a-4c7e-a646-047bcfd3046f-openstack-config\") pod \"openstackclient\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " pod="openstack/openstackclient" Dec 01 22:27:00 crc kubenswrapper[4857]: I1201 22:27:00.996557 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f50f03d-e84a-4c7e-a646-047bcfd3046f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " pod="openstack/openstackclient" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.097750 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49nzh\" (UniqueName: \"kubernetes.io/projected/2f50f03d-e84a-4c7e-a646-047bcfd3046f-kube-api-access-49nzh\") pod \"openstackclient\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " pod="openstack/openstackclient" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.097854 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2f50f03d-e84a-4c7e-a646-047bcfd3046f-openstack-config-secret\") pod \"openstackclient\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " pod="openstack/openstackclient" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.097905 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2f50f03d-e84a-4c7e-a646-047bcfd3046f-openstack-config\") pod \"openstackclient\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " pod="openstack/openstackclient" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.097936 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f50f03d-e84a-4c7e-a646-047bcfd3046f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " pod="openstack/openstackclient" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.098917 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2f50f03d-e84a-4c7e-a646-047bcfd3046f-openstack-config\") pod \"openstackclient\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " pod="openstack/openstackclient" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.103510 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2f50f03d-e84a-4c7e-a646-047bcfd3046f-openstack-config-secret\") pod \"openstackclient\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " pod="openstack/openstackclient" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.105639 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f50f03d-e84a-4c7e-a646-047bcfd3046f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " pod="openstack/openstackclient" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.124951 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49nzh\" (UniqueName: \"kubernetes.io/projected/2f50f03d-e84a-4c7e-a646-047bcfd3046f-kube-api-access-49nzh\") pod \"openstackclient\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " pod="openstack/openstackclient" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.206075 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.209675 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="4ea60739-fcd8-45c6-9353-b44dd058f404" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.213394 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.218639 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.221245 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="4ea60739-fcd8-45c6-9353-b44dd058f404" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.743822 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.849143 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ea60739-fcd8-45c6-9353-b44dd058f404" path="/var/lib/kubelet/pods/4ea60739-fcd8-45c6-9353-b44dd058f404/volumes" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.944328 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-mh9p4"] Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.945692 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-mh9p4" Dec 01 22:27:01 crc kubenswrapper[4857]: I1201 22:27:01.963769 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-mh9p4"] Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.042391 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-487f-account-create-update-vszxx"] Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.043719 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-487f-account-create-update-vszxx" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.045902 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.059104 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-487f-account-create-update-vszxx"] Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.118358 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e58803a1-45fc-4be0-9f0e-536338384e04-operator-scripts\") pod \"aodh-db-create-mh9p4\" (UID: \"e58803a1-45fc-4be0-9f0e-536338384e04\") " pod="openstack/aodh-db-create-mh9p4" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.119258 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7scqh\" (UniqueName: \"kubernetes.io/projected/e58803a1-45fc-4be0-9f0e-536338384e04-kube-api-access-7scqh\") pod \"aodh-db-create-mh9p4\" (UID: \"e58803a1-45fc-4be0-9f0e-536338384e04\") " pod="openstack/aodh-db-create-mh9p4" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.215422 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.215420 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2f50f03d-e84a-4c7e-a646-047bcfd3046f","Type":"ContainerStarted","Data":"134f2bba3e91ccb269c800ef4cdd79e63cf2a734297141290056902d1d9e4c57"} Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.215468 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2f50f03d-e84a-4c7e-a646-047bcfd3046f","Type":"ContainerStarted","Data":"659c1811ac9c8aa5a4f49e0d1537fdc4c207f39924653c1e4d42428621b0f70e"} Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.218741 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="4ea60739-fcd8-45c6-9353-b44dd058f404" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.220562 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0589b0a-71fe-47a1-9dd3-981c43ac56f5-operator-scripts\") pod \"aodh-487f-account-create-update-vszxx\" (UID: \"f0589b0a-71fe-47a1-9dd3-981c43ac56f5\") " pod="openstack/aodh-487f-account-create-update-vszxx" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.221182 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7g2z\" (UniqueName: \"kubernetes.io/projected/f0589b0a-71fe-47a1-9dd3-981c43ac56f5-kube-api-access-w7g2z\") pod \"aodh-487f-account-create-update-vszxx\" (UID: \"f0589b0a-71fe-47a1-9dd3-981c43ac56f5\") " pod="openstack/aodh-487f-account-create-update-vszxx" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.221278 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e58803a1-45fc-4be0-9f0e-536338384e04-operator-scripts\") pod \"aodh-db-create-mh9p4\" (UID: \"e58803a1-45fc-4be0-9f0e-536338384e04\") " pod="openstack/aodh-db-create-mh9p4" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.221614 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7scqh\" (UniqueName: \"kubernetes.io/projected/e58803a1-45fc-4be0-9f0e-536338384e04-kube-api-access-7scqh\") pod \"aodh-db-create-mh9p4\" (UID: \"e58803a1-45fc-4be0-9f0e-536338384e04\") " pod="openstack/aodh-db-create-mh9p4" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.221967 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e58803a1-45fc-4be0-9f0e-536338384e04-operator-scripts\") pod \"aodh-db-create-mh9p4\" (UID: \"e58803a1-45fc-4be0-9f0e-536338384e04\") " pod="openstack/aodh-db-create-mh9p4" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.242659 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7scqh\" (UniqueName: \"kubernetes.io/projected/e58803a1-45fc-4be0-9f0e-536338384e04-kube-api-access-7scqh\") pod \"aodh-db-create-mh9p4\" (UID: \"e58803a1-45fc-4be0-9f0e-536338384e04\") " pod="openstack/aodh-db-create-mh9p4" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.246610 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="4ea60739-fcd8-45c6-9353-b44dd058f404" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.250365 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.2503451930000002 podStartE2EDuration="2.250345193s" podCreationTimestamp="2025-12-01 22:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 22:27:02.238410927 +0000 UTC m=+3140.728473274" watchObservedRunningTime="2025-12-01 22:27:02.250345193 +0000 UTC m=+3140.740407510" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.259947 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-mh9p4" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.323995 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0589b0a-71fe-47a1-9dd3-981c43ac56f5-operator-scripts\") pod \"aodh-487f-account-create-update-vszxx\" (UID: \"f0589b0a-71fe-47a1-9dd3-981c43ac56f5\") " pod="openstack/aodh-487f-account-create-update-vszxx" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.324150 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7g2z\" (UniqueName: \"kubernetes.io/projected/f0589b0a-71fe-47a1-9dd3-981c43ac56f5-kube-api-access-w7g2z\") pod \"aodh-487f-account-create-update-vszxx\" (UID: \"f0589b0a-71fe-47a1-9dd3-981c43ac56f5\") " pod="openstack/aodh-487f-account-create-update-vszxx" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.324829 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0589b0a-71fe-47a1-9dd3-981c43ac56f5-operator-scripts\") pod \"aodh-487f-account-create-update-vszxx\" (UID: \"f0589b0a-71fe-47a1-9dd3-981c43ac56f5\") " pod="openstack/aodh-487f-account-create-update-vszxx" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.342396 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7g2z\" (UniqueName: \"kubernetes.io/projected/f0589b0a-71fe-47a1-9dd3-981c43ac56f5-kube-api-access-w7g2z\") pod \"aodh-487f-account-create-update-vszxx\" (UID: \"f0589b0a-71fe-47a1-9dd3-981c43ac56f5\") " pod="openstack/aodh-487f-account-create-update-vszxx" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.364140 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-487f-account-create-update-vszxx" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.742796 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-mh9p4"] Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.849395 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-487f-account-create-update-vszxx"] Dec 01 22:27:02 crc kubenswrapper[4857]: W1201 22:27:02.856422 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0589b0a_71fe_47a1_9dd3_981c43ac56f5.slice/crio-7d7fb611390aed1804d238b478019cd250eb9fa9e8305f3502817d3e81eef8b7 WatchSource:0}: Error finding container 7d7fb611390aed1804d238b478019cd250eb9fa9e8305f3502817d3e81eef8b7: Status 404 returned error can't find the container with id 7d7fb611390aed1804d238b478019cd250eb9fa9e8305f3502817d3e81eef8b7 Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.905821 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.912219 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="6821156d-2d04-4608-a5c7-c70b1058fdc0" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.939293 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6821156d-2d04-4608-a5c7-c70b1058fdc0-openstack-config-secret\") pod \"6821156d-2d04-4608-a5c7-c70b1058fdc0\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.939375 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfq9q\" (UniqueName: \"kubernetes.io/projected/6821156d-2d04-4608-a5c7-c70b1058fdc0-kube-api-access-bfq9q\") pod \"6821156d-2d04-4608-a5c7-c70b1058fdc0\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.939490 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6821156d-2d04-4608-a5c7-c70b1058fdc0-openstack-config\") pod \"6821156d-2d04-4608-a5c7-c70b1058fdc0\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.939591 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6821156d-2d04-4608-a5c7-c70b1058fdc0-combined-ca-bundle\") pod \"6821156d-2d04-4608-a5c7-c70b1058fdc0\" (UID: \"6821156d-2d04-4608-a5c7-c70b1058fdc0\") " Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.946150 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6821156d-2d04-4608-a5c7-c70b1058fdc0-kube-api-access-bfq9q" (OuterVolumeSpecName: "kube-api-access-bfq9q") pod "6821156d-2d04-4608-a5c7-c70b1058fdc0" (UID: "6821156d-2d04-4608-a5c7-c70b1058fdc0"). InnerVolumeSpecName "kube-api-access-bfq9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.970932 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6821156d-2d04-4608-a5c7-c70b1058fdc0-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "6821156d-2d04-4608-a5c7-c70b1058fdc0" (UID: "6821156d-2d04-4608-a5c7-c70b1058fdc0"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:27:02 crc kubenswrapper[4857]: I1201 22:27:02.997959 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6821156d-2d04-4608-a5c7-c70b1058fdc0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6821156d-2d04-4608-a5c7-c70b1058fdc0" (UID: "6821156d-2d04-4608-a5c7-c70b1058fdc0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.009926 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6821156d-2d04-4608-a5c7-c70b1058fdc0-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "6821156d-2d04-4608-a5c7-c70b1058fdc0" (UID: "6821156d-2d04-4608-a5c7-c70b1058fdc0"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.042211 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfq9q\" (UniqueName: \"kubernetes.io/projected/6821156d-2d04-4608-a5c7-c70b1058fdc0-kube-api-access-bfq9q\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.042240 4857 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6821156d-2d04-4608-a5c7-c70b1058fdc0-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.042249 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6821156d-2d04-4608-a5c7-c70b1058fdc0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.042258 4857 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6821156d-2d04-4608-a5c7-c70b1058fdc0-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.223717 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-487f-account-create-update-vszxx" event={"ID":"f0589b0a-71fe-47a1-9dd3-981c43ac56f5","Type":"ContainerStarted","Data":"b24a0b8d6a1bb24674b263d6cc883d208a602e5c498f1252f0d2e318065873ce"} Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.223760 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-487f-account-create-update-vszxx" event={"ID":"f0589b0a-71fe-47a1-9dd3-981c43ac56f5","Type":"ContainerStarted","Data":"7d7fb611390aed1804d238b478019cd250eb9fa9e8305f3502817d3e81eef8b7"} Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.225443 4857 generic.go:334] "Generic (PLEG): container finished" podID="e58803a1-45fc-4be0-9f0e-536338384e04" containerID="66ad3aa84195ee062595491b192c68262235c71f3a10bbfcb59c051fa27717e3" exitCode=0 Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.225542 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-mh9p4" event={"ID":"e58803a1-45fc-4be0-9f0e-536338384e04","Type":"ContainerDied","Data":"66ad3aa84195ee062595491b192c68262235c71f3a10bbfcb59c051fa27717e3"} Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.225598 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-mh9p4" event={"ID":"e58803a1-45fc-4be0-9f0e-536338384e04","Type":"ContainerStarted","Data":"9ac5b350a34ad4d545edb9d26b0db2109d0c65fc857acf207fda4907fc9c8052"} Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.226799 4857 generic.go:334] "Generic (PLEG): container finished" podID="6821156d-2d04-4608-a5c7-c70b1058fdc0" containerID="c657db47da331af1813f6ade9364da74d4954e848ff4cf4cb00cf13198f57d1d" exitCode=137 Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.227297 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.229137 4857 scope.go:117] "RemoveContainer" containerID="c657db47da331af1813f6ade9364da74d4954e848ff4cf4cb00cf13198f57d1d" Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.246814 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-487f-account-create-update-vszxx" podStartSLOduration=1.246796219 podStartE2EDuration="1.246796219s" podCreationTimestamp="2025-12-01 22:27:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 22:27:03.24480974 +0000 UTC m=+3141.734872057" watchObservedRunningTime="2025-12-01 22:27:03.246796219 +0000 UTC m=+3141.736858536" Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.247435 4857 scope.go:117] "RemoveContainer" containerID="c657db47da331af1813f6ade9364da74d4954e848ff4cf4cb00cf13198f57d1d" Dec 01 22:27:03 crc kubenswrapper[4857]: E1201 22:27:03.247980 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c657db47da331af1813f6ade9364da74d4954e848ff4cf4cb00cf13198f57d1d\": container with ID starting with c657db47da331af1813f6ade9364da74d4954e848ff4cf4cb00cf13198f57d1d not found: ID does not exist" containerID="c657db47da331af1813f6ade9364da74d4954e848ff4cf4cb00cf13198f57d1d" Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.248021 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c657db47da331af1813f6ade9364da74d4954e848ff4cf4cb00cf13198f57d1d"} err="failed to get container status \"c657db47da331af1813f6ade9364da74d4954e848ff4cf4cb00cf13198f57d1d\": rpc error: code = NotFound desc = could not find container \"c657db47da331af1813f6ade9364da74d4954e848ff4cf4cb00cf13198f57d1d\": container with ID starting with c657db47da331af1813f6ade9364da74d4954e848ff4cf4cb00cf13198f57d1d not found: ID does not exist" Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.264873 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="6821156d-2d04-4608-a5c7-c70b1058fdc0" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" Dec 01 22:27:03 crc kubenswrapper[4857]: I1201 22:27:03.847131 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6821156d-2d04-4608-a5c7-c70b1058fdc0" path="/var/lib/kubelet/pods/6821156d-2d04-4608-a5c7-c70b1058fdc0/volumes" Dec 01 22:27:04 crc kubenswrapper[4857]: I1201 22:27:04.240857 4857 generic.go:334] "Generic (PLEG): container finished" podID="f0589b0a-71fe-47a1-9dd3-981c43ac56f5" containerID="b24a0b8d6a1bb24674b263d6cc883d208a602e5c498f1252f0d2e318065873ce" exitCode=0 Dec 01 22:27:04 crc kubenswrapper[4857]: I1201 22:27:04.240978 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-487f-account-create-update-vszxx" event={"ID":"f0589b0a-71fe-47a1-9dd3-981c43ac56f5","Type":"ContainerDied","Data":"b24a0b8d6a1bb24674b263d6cc883d208a602e5c498f1252f0d2e318065873ce"} Dec 01 22:27:04 crc kubenswrapper[4857]: I1201 22:27:04.684486 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-mh9p4" Dec 01 22:27:04 crc kubenswrapper[4857]: I1201 22:27:04.781165 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e58803a1-45fc-4be0-9f0e-536338384e04-operator-scripts\") pod \"e58803a1-45fc-4be0-9f0e-536338384e04\" (UID: \"e58803a1-45fc-4be0-9f0e-536338384e04\") " Dec 01 22:27:04 crc kubenswrapper[4857]: I1201 22:27:04.781264 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7scqh\" (UniqueName: \"kubernetes.io/projected/e58803a1-45fc-4be0-9f0e-536338384e04-kube-api-access-7scqh\") pod \"e58803a1-45fc-4be0-9f0e-536338384e04\" (UID: \"e58803a1-45fc-4be0-9f0e-536338384e04\") " Dec 01 22:27:04 crc kubenswrapper[4857]: I1201 22:27:04.782293 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e58803a1-45fc-4be0-9f0e-536338384e04-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e58803a1-45fc-4be0-9f0e-536338384e04" (UID: "e58803a1-45fc-4be0-9f0e-536338384e04"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:27:04 crc kubenswrapper[4857]: I1201 22:27:04.786944 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e58803a1-45fc-4be0-9f0e-536338384e04-kube-api-access-7scqh" (OuterVolumeSpecName: "kube-api-access-7scqh") pod "e58803a1-45fc-4be0-9f0e-536338384e04" (UID: "e58803a1-45fc-4be0-9f0e-536338384e04"). InnerVolumeSpecName "kube-api-access-7scqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:27:04 crc kubenswrapper[4857]: I1201 22:27:04.883031 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e58803a1-45fc-4be0-9f0e-536338384e04-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:04 crc kubenswrapper[4857]: I1201 22:27:04.883084 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7scqh\" (UniqueName: \"kubernetes.io/projected/e58803a1-45fc-4be0-9f0e-536338384e04-kube-api-access-7scqh\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:05 crc kubenswrapper[4857]: I1201 22:27:05.258154 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-mh9p4" event={"ID":"e58803a1-45fc-4be0-9f0e-536338384e04","Type":"ContainerDied","Data":"9ac5b350a34ad4d545edb9d26b0db2109d0c65fc857acf207fda4907fc9c8052"} Dec 01 22:27:05 crc kubenswrapper[4857]: I1201 22:27:05.258195 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-mh9p4" Dec 01 22:27:05 crc kubenswrapper[4857]: I1201 22:27:05.258212 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ac5b350a34ad4d545edb9d26b0db2109d0c65fc857acf207fda4907fc9c8052" Dec 01 22:27:05 crc kubenswrapper[4857]: I1201 22:27:05.748646 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-487f-account-create-update-vszxx" Dec 01 22:27:05 crc kubenswrapper[4857]: I1201 22:27:05.804876 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0589b0a-71fe-47a1-9dd3-981c43ac56f5-operator-scripts\") pod \"f0589b0a-71fe-47a1-9dd3-981c43ac56f5\" (UID: \"f0589b0a-71fe-47a1-9dd3-981c43ac56f5\") " Dec 01 22:27:05 crc kubenswrapper[4857]: I1201 22:27:05.805201 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7g2z\" (UniqueName: \"kubernetes.io/projected/f0589b0a-71fe-47a1-9dd3-981c43ac56f5-kube-api-access-w7g2z\") pod \"f0589b0a-71fe-47a1-9dd3-981c43ac56f5\" (UID: \"f0589b0a-71fe-47a1-9dd3-981c43ac56f5\") " Dec 01 22:27:05 crc kubenswrapper[4857]: I1201 22:27:05.805480 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0589b0a-71fe-47a1-9dd3-981c43ac56f5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f0589b0a-71fe-47a1-9dd3-981c43ac56f5" (UID: "f0589b0a-71fe-47a1-9dd3-981c43ac56f5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:27:05 crc kubenswrapper[4857]: I1201 22:27:05.805865 4857 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0589b0a-71fe-47a1-9dd3-981c43ac56f5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:05 crc kubenswrapper[4857]: I1201 22:27:05.822473 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0589b0a-71fe-47a1-9dd3-981c43ac56f5-kube-api-access-w7g2z" (OuterVolumeSpecName: "kube-api-access-w7g2z") pod "f0589b0a-71fe-47a1-9dd3-981c43ac56f5" (UID: "f0589b0a-71fe-47a1-9dd3-981c43ac56f5"). InnerVolumeSpecName "kube-api-access-w7g2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:27:05 crc kubenswrapper[4857]: I1201 22:27:05.907838 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7g2z\" (UniqueName: \"kubernetes.io/projected/f0589b0a-71fe-47a1-9dd3-981c43ac56f5-kube-api-access-w7g2z\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:06 crc kubenswrapper[4857]: I1201 22:27:06.270064 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-487f-account-create-update-vszxx" event={"ID":"f0589b0a-71fe-47a1-9dd3-981c43ac56f5","Type":"ContainerDied","Data":"7d7fb611390aed1804d238b478019cd250eb9fa9e8305f3502817d3e81eef8b7"} Dec 01 22:27:06 crc kubenswrapper[4857]: I1201 22:27:06.270105 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d7fb611390aed1804d238b478019cd250eb9fa9e8305f3502817d3e81eef8b7" Dec 01 22:27:06 crc kubenswrapper[4857]: I1201 22:27:06.270170 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-487f-account-create-update-vszxx" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.446548 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-k85cf"] Dec 01 22:27:07 crc kubenswrapper[4857]: E1201 22:27:07.447370 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e58803a1-45fc-4be0-9f0e-536338384e04" containerName="mariadb-database-create" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.447388 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e58803a1-45fc-4be0-9f0e-536338384e04" containerName="mariadb-database-create" Dec 01 22:27:07 crc kubenswrapper[4857]: E1201 22:27:07.447426 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0589b0a-71fe-47a1-9dd3-981c43ac56f5" containerName="mariadb-account-create-update" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.447438 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0589b0a-71fe-47a1-9dd3-981c43ac56f5" containerName="mariadb-account-create-update" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.447689 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e58803a1-45fc-4be0-9f0e-536338384e04" containerName="mariadb-database-create" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.447715 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0589b0a-71fe-47a1-9dd3-981c43ac56f5" containerName="mariadb-account-create-update" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.448542 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.453142 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-j6fzf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.453256 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.453405 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.458289 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.460489 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-k85cf"] Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.491098 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-config-data\") pod \"aodh-db-sync-k85cf\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.491194 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-scripts\") pod \"aodh-db-sync-k85cf\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.491246 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-combined-ca-bundle\") pod \"aodh-db-sync-k85cf\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.491269 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fksv\" (UniqueName: \"kubernetes.io/projected/0bf9c958-5340-4c35-a0d5-fc5693c9f589-kube-api-access-9fksv\") pod \"aodh-db-sync-k85cf\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.592857 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-config-data\") pod \"aodh-db-sync-k85cf\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.592907 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-scripts\") pod \"aodh-db-sync-k85cf\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.592938 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-combined-ca-bundle\") pod \"aodh-db-sync-k85cf\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.592962 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fksv\" (UniqueName: \"kubernetes.io/projected/0bf9c958-5340-4c35-a0d5-fc5693c9f589-kube-api-access-9fksv\") pod \"aodh-db-sync-k85cf\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.597219 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-combined-ca-bundle\") pod \"aodh-db-sync-k85cf\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.599693 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-config-data\") pod \"aodh-db-sync-k85cf\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.602370 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-scripts\") pod \"aodh-db-sync-k85cf\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.627526 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fksv\" (UniqueName: \"kubernetes.io/projected/0bf9c958-5340-4c35-a0d5-fc5693c9f589-kube-api-access-9fksv\") pod \"aodh-db-sync-k85cf\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:07 crc kubenswrapper[4857]: I1201 22:27:07.781444 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:08 crc kubenswrapper[4857]: I1201 22:27:08.388418 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-k85cf"] Dec 01 22:27:08 crc kubenswrapper[4857]: W1201 22:27:08.396468 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0bf9c958_5340_4c35_a0d5_fc5693c9f589.slice/crio-edc0ce39376834da1aa52c9ca65a9b9cb864b8606085c892e1d756b7f98a18ad WatchSource:0}: Error finding container edc0ce39376834da1aa52c9ca65a9b9cb864b8606085c892e1d756b7f98a18ad: Status 404 returned error can't find the container with id edc0ce39376834da1aa52c9ca65a9b9cb864b8606085c892e1d756b7f98a18ad Dec 01 22:27:08 crc kubenswrapper[4857]: I1201 22:27:08.835347 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:27:08 crc kubenswrapper[4857]: E1201 22:27:08.835764 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:27:09 crc kubenswrapper[4857]: I1201 22:27:09.307447 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-k85cf" event={"ID":"0bf9c958-5340-4c35-a0d5-fc5693c9f589","Type":"ContainerStarted","Data":"edc0ce39376834da1aa52c9ca65a9b9cb864b8606085c892e1d756b7f98a18ad"} Dec 01 22:27:14 crc kubenswrapper[4857]: I1201 22:27:14.361711 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-k85cf" event={"ID":"0bf9c958-5340-4c35-a0d5-fc5693c9f589","Type":"ContainerStarted","Data":"ce32b9add64c8ccbfc5d066896bab28739536bc89824ebc0cf77cae4ecc7e717"} Dec 01 22:27:14 crc kubenswrapper[4857]: I1201 22:27:14.399088 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-k85cf" podStartSLOduration=2.320206811 podStartE2EDuration="7.39902942s" podCreationTimestamp="2025-12-01 22:27:07 +0000 UTC" firstStartedPulling="2025-12-01 22:27:08.404481824 +0000 UTC m=+3146.894544151" lastFinishedPulling="2025-12-01 22:27:13.483304443 +0000 UTC m=+3151.973366760" observedRunningTime="2025-12-01 22:27:14.384572722 +0000 UTC m=+3152.874635069" watchObservedRunningTime="2025-12-01 22:27:14.39902942 +0000 UTC m=+3152.889091767" Dec 01 22:27:16 crc kubenswrapper[4857]: I1201 22:27:16.384819 4857 generic.go:334] "Generic (PLEG): container finished" podID="0bf9c958-5340-4c35-a0d5-fc5693c9f589" containerID="ce32b9add64c8ccbfc5d066896bab28739536bc89824ebc0cf77cae4ecc7e717" exitCode=0 Dec 01 22:27:16 crc kubenswrapper[4857]: I1201 22:27:16.384918 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-k85cf" event={"ID":"0bf9c958-5340-4c35-a0d5-fc5693c9f589","Type":"ContainerDied","Data":"ce32b9add64c8ccbfc5d066896bab28739536bc89824ebc0cf77cae4ecc7e717"} Dec 01 22:27:17 crc kubenswrapper[4857]: I1201 22:27:17.809269 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:17 crc kubenswrapper[4857]: I1201 22:27:17.965616 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fksv\" (UniqueName: \"kubernetes.io/projected/0bf9c958-5340-4c35-a0d5-fc5693c9f589-kube-api-access-9fksv\") pod \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " Dec 01 22:27:17 crc kubenswrapper[4857]: I1201 22:27:17.965709 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-config-data\") pod \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " Dec 01 22:27:17 crc kubenswrapper[4857]: I1201 22:27:17.965753 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-combined-ca-bundle\") pod \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " Dec 01 22:27:17 crc kubenswrapper[4857]: I1201 22:27:17.965820 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-scripts\") pod \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\" (UID: \"0bf9c958-5340-4c35-a0d5-fc5693c9f589\") " Dec 01 22:27:17 crc kubenswrapper[4857]: I1201 22:27:17.971540 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf9c958-5340-4c35-a0d5-fc5693c9f589-kube-api-access-9fksv" (OuterVolumeSpecName: "kube-api-access-9fksv") pod "0bf9c958-5340-4c35-a0d5-fc5693c9f589" (UID: "0bf9c958-5340-4c35-a0d5-fc5693c9f589"). InnerVolumeSpecName "kube-api-access-9fksv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:27:17 crc kubenswrapper[4857]: I1201 22:27:17.977300 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-scripts" (OuterVolumeSpecName: "scripts") pod "0bf9c958-5340-4c35-a0d5-fc5693c9f589" (UID: "0bf9c958-5340-4c35-a0d5-fc5693c9f589"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:27:18 crc kubenswrapper[4857]: I1201 22:27:18.005293 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-config-data" (OuterVolumeSpecName: "config-data") pod "0bf9c958-5340-4c35-a0d5-fc5693c9f589" (UID: "0bf9c958-5340-4c35-a0d5-fc5693c9f589"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:27:18 crc kubenswrapper[4857]: I1201 22:27:18.030364 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0bf9c958-5340-4c35-a0d5-fc5693c9f589" (UID: "0bf9c958-5340-4c35-a0d5-fc5693c9f589"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:27:18 crc kubenswrapper[4857]: I1201 22:27:18.068159 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fksv\" (UniqueName: \"kubernetes.io/projected/0bf9c958-5340-4c35-a0d5-fc5693c9f589-kube-api-access-9fksv\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:18 crc kubenswrapper[4857]: I1201 22:27:18.068192 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:18 crc kubenswrapper[4857]: I1201 22:27:18.068206 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:18 crc kubenswrapper[4857]: I1201 22:27:18.068217 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf9c958-5340-4c35-a0d5-fc5693c9f589-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:18 crc kubenswrapper[4857]: I1201 22:27:18.437374 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-k85cf" event={"ID":"0bf9c958-5340-4c35-a0d5-fc5693c9f589","Type":"ContainerDied","Data":"edc0ce39376834da1aa52c9ca65a9b9cb864b8606085c892e1d756b7f98a18ad"} Dec 01 22:27:18 crc kubenswrapper[4857]: I1201 22:27:18.437739 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edc0ce39376834da1aa52c9ca65a9b9cb864b8606085c892e1d756b7f98a18ad" Dec 01 22:27:18 crc kubenswrapper[4857]: I1201 22:27:18.437867 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-k85cf" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.551126 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 01 22:27:22 crc kubenswrapper[4857]: E1201 22:27:22.552558 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf9c958-5340-4c35-a0d5-fc5693c9f589" containerName="aodh-db-sync" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.552585 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf9c958-5340-4c35-a0d5-fc5693c9f589" containerName="aodh-db-sync" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.552952 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf9c958-5340-4c35-a0d5-fc5693c9f589" containerName="aodh-db-sync" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.556214 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.560103 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-j6fzf" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.560394 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.560613 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.563290 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.703848 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-combined-ca-bundle\") pod \"aodh-0\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.704193 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb2sz\" (UniqueName: \"kubernetes.io/projected/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-kube-api-access-gb2sz\") pod \"aodh-0\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.704220 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-config-data\") pod \"aodh-0\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.704251 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-scripts\") pod \"aodh-0\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.805617 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-combined-ca-bundle\") pod \"aodh-0\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.805918 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb2sz\" (UniqueName: \"kubernetes.io/projected/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-kube-api-access-gb2sz\") pod \"aodh-0\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.806015 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-config-data\") pod \"aodh-0\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.806151 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-scripts\") pod \"aodh-0\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.818868 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-combined-ca-bundle\") pod \"aodh-0\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.819297 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-scripts\") pod \"aodh-0\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.820144 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-config-data\") pod \"aodh-0\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.824532 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb2sz\" (UniqueName: \"kubernetes.io/projected/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-kube-api-access-gb2sz\") pod \"aodh-0\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " pod="openstack/aodh-0" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.835082 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:27:22 crc kubenswrapper[4857]: E1201 22:27:22.835371 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:27:22 crc kubenswrapper[4857]: I1201 22:27:22.920903 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:27:23 crc kubenswrapper[4857]: I1201 22:27:23.418564 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 22:27:23 crc kubenswrapper[4857]: I1201 22:27:23.491395 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"eaa5cc4c-c065-45ab-a735-2dfd6af03afe","Type":"ContainerStarted","Data":"225850ee9a1a34b0e9b00b3f5c91987cb07de39461dbf1501bb53e5f2aa04bb4"} Dec 01 22:27:24 crc kubenswrapper[4857]: I1201 22:27:24.523536 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 22:27:24 crc kubenswrapper[4857]: I1201 22:27:24.524019 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="ceilometer-central-agent" containerID="cri-o://85615bd37db3ac0824557104e6025e64e6f8570c228d3a62f90bb002e61e6d86" gracePeriod=30 Dec 01 22:27:24 crc kubenswrapper[4857]: I1201 22:27:24.524158 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="proxy-httpd" containerID="cri-o://6e2ea6014b27950b384592d210206391045a8404ba8de5cda7ffeacd1e103116" gracePeriod=30 Dec 01 22:27:24 crc kubenswrapper[4857]: I1201 22:27:24.524201 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="sg-core" containerID="cri-o://cb2192707cd1882c643688790bbc8111c00787d69d16d549232bd81f26c9e7d8" gracePeriod=30 Dec 01 22:27:24 crc kubenswrapper[4857]: I1201 22:27:24.524233 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="ceilometer-notification-agent" containerID="cri-o://1953a1657ee82da7d20d764e11157274898e04446c6a3dd4976b6233c0e80149" gracePeriod=30 Dec 01 22:27:25 crc kubenswrapper[4857]: I1201 22:27:25.512573 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"eaa5cc4c-c065-45ab-a735-2dfd6af03afe","Type":"ContainerStarted","Data":"996fefaac23ac63ff0afd65bea71df4c08d41a78268be04474cc652807258715"} Dec 01 22:27:25 crc kubenswrapper[4857]: I1201 22:27:25.514974 4857 generic.go:334] "Generic (PLEG): container finished" podID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerID="6e2ea6014b27950b384592d210206391045a8404ba8de5cda7ffeacd1e103116" exitCode=0 Dec 01 22:27:25 crc kubenswrapper[4857]: I1201 22:27:25.514994 4857 generic.go:334] "Generic (PLEG): container finished" podID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerID="cb2192707cd1882c643688790bbc8111c00787d69d16d549232bd81f26c9e7d8" exitCode=2 Dec 01 22:27:25 crc kubenswrapper[4857]: I1201 22:27:25.515001 4857 generic.go:334] "Generic (PLEG): container finished" podID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerID="85615bd37db3ac0824557104e6025e64e6f8570c228d3a62f90bb002e61e6d86" exitCode=0 Dec 01 22:27:25 crc kubenswrapper[4857]: I1201 22:27:25.515015 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed82b79a-baa2-42ed-a35c-935dfacc778f","Type":"ContainerDied","Data":"6e2ea6014b27950b384592d210206391045a8404ba8de5cda7ffeacd1e103116"} Dec 01 22:27:25 crc kubenswrapper[4857]: I1201 22:27:25.515028 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed82b79a-baa2-42ed-a35c-935dfacc778f","Type":"ContainerDied","Data":"cb2192707cd1882c643688790bbc8111c00787d69d16d549232bd81f26c9e7d8"} Dec 01 22:27:25 crc kubenswrapper[4857]: I1201 22:27:25.515051 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed82b79a-baa2-42ed-a35c-935dfacc778f","Type":"ContainerDied","Data":"85615bd37db3ac0824557104e6025e64e6f8570c228d3a62f90bb002e61e6d86"} Dec 01 22:27:25 crc kubenswrapper[4857]: I1201 22:27:25.821428 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 22:27:26 crc kubenswrapper[4857]: I1201 22:27:26.524135 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"eaa5cc4c-c065-45ab-a735-2dfd6af03afe","Type":"ContainerStarted","Data":"06ce657f9604cee3e2f96b8b52ced69d57105d8a456835960424448358c665fc"} Dec 01 22:27:28 crc kubenswrapper[4857]: I1201 22:27:28.558758 4857 generic.go:334] "Generic (PLEG): container finished" podID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerID="1953a1657ee82da7d20d764e11157274898e04446c6a3dd4976b6233c0e80149" exitCode=0 Dec 01 22:27:28 crc kubenswrapper[4857]: I1201 22:27:28.559304 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed82b79a-baa2-42ed-a35c-935dfacc778f","Type":"ContainerDied","Data":"1953a1657ee82da7d20d764e11157274898e04446c6a3dd4976b6233c0e80149"} Dec 01 22:27:28 crc kubenswrapper[4857]: I1201 22:27:28.560998 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"eaa5cc4c-c065-45ab-a735-2dfd6af03afe","Type":"ContainerStarted","Data":"4702163fa4c65d550ec70cb4b1c15e4c93137d835ae245ef328119d26963f795"} Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.342140 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.444878 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-ceilometer-tls-certs\") pod \"ed82b79a-baa2-42ed-a35c-935dfacc778f\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.444939 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed82b79a-baa2-42ed-a35c-935dfacc778f-run-httpd\") pod \"ed82b79a-baa2-42ed-a35c-935dfacc778f\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.444993 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-scripts\") pod \"ed82b79a-baa2-42ed-a35c-935dfacc778f\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.445014 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-config-data\") pod \"ed82b79a-baa2-42ed-a35c-935dfacc778f\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.445063 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-combined-ca-bundle\") pod \"ed82b79a-baa2-42ed-a35c-935dfacc778f\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.445125 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-sg-core-conf-yaml\") pod \"ed82b79a-baa2-42ed-a35c-935dfacc778f\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.445155 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74dws\" (UniqueName: \"kubernetes.io/projected/ed82b79a-baa2-42ed-a35c-935dfacc778f-kube-api-access-74dws\") pod \"ed82b79a-baa2-42ed-a35c-935dfacc778f\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.445193 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed82b79a-baa2-42ed-a35c-935dfacc778f-log-httpd\") pod \"ed82b79a-baa2-42ed-a35c-935dfacc778f\" (UID: \"ed82b79a-baa2-42ed-a35c-935dfacc778f\") " Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.446362 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed82b79a-baa2-42ed-a35c-935dfacc778f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ed82b79a-baa2-42ed-a35c-935dfacc778f" (UID: "ed82b79a-baa2-42ed-a35c-935dfacc778f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.446689 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed82b79a-baa2-42ed-a35c-935dfacc778f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ed82b79a-baa2-42ed-a35c-935dfacc778f" (UID: "ed82b79a-baa2-42ed-a35c-935dfacc778f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.454347 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-scripts" (OuterVolumeSpecName: "scripts") pod "ed82b79a-baa2-42ed-a35c-935dfacc778f" (UID: "ed82b79a-baa2-42ed-a35c-935dfacc778f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.466116 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed82b79a-baa2-42ed-a35c-935dfacc778f-kube-api-access-74dws" (OuterVolumeSpecName: "kube-api-access-74dws") pod "ed82b79a-baa2-42ed-a35c-935dfacc778f" (UID: "ed82b79a-baa2-42ed-a35c-935dfacc778f"). InnerVolumeSpecName "kube-api-access-74dws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.474951 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ed82b79a-baa2-42ed-a35c-935dfacc778f" (UID: "ed82b79a-baa2-42ed-a35c-935dfacc778f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.528988 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ed82b79a-baa2-42ed-a35c-935dfacc778f" (UID: "ed82b79a-baa2-42ed-a35c-935dfacc778f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.550269 4857 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.550304 4857 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed82b79a-baa2-42ed-a35c-935dfacc778f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.550320 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.550333 4857 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.550345 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74dws\" (UniqueName: \"kubernetes.io/projected/ed82b79a-baa2-42ed-a35c-935dfacc778f-kube-api-access-74dws\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.550359 4857 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed82b79a-baa2-42ed-a35c-935dfacc778f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.560034 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed82b79a-baa2-42ed-a35c-935dfacc778f" (UID: "ed82b79a-baa2-42ed-a35c-935dfacc778f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.573966 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"eaa5cc4c-c065-45ab-a735-2dfd6af03afe","Type":"ContainerStarted","Data":"d39378bf79c21b3adfa86e270c6cd68b9434e065b3e0a66169d33e26935bd602"} Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.574169 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-api" containerID="cri-o://996fefaac23ac63ff0afd65bea71df4c08d41a78268be04474cc652807258715" gracePeriod=30 Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.574600 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-notifier" containerID="cri-o://4702163fa4c65d550ec70cb4b1c15e4c93137d835ae245ef328119d26963f795" gracePeriod=30 Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.574663 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-listener" containerID="cri-o://d39378bf79c21b3adfa86e270c6cd68b9434e065b3e0a66169d33e26935bd602" gracePeriod=30 Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.574763 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-evaluator" containerID="cri-o://06ce657f9604cee3e2f96b8b52ced69d57105d8a456835960424448358c665fc" gracePeriod=30 Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.581494 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed82b79a-baa2-42ed-a35c-935dfacc778f","Type":"ContainerDied","Data":"3d447f8817e1f4b21e0e8c51cdcdcb49eb15cf8a0ad37df746ad030807b3b820"} Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.581550 4857 scope.go:117] "RemoveContainer" containerID="6e2ea6014b27950b384592d210206391045a8404ba8de5cda7ffeacd1e103116" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.581558 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.583134 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-config-data" (OuterVolumeSpecName: "config-data") pod "ed82b79a-baa2-42ed-a35c-935dfacc778f" (UID: "ed82b79a-baa2-42ed-a35c-935dfacc778f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.600233 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.847399687 podStartE2EDuration="7.600209795s" podCreationTimestamp="2025-12-01 22:27:22 +0000 UTC" firstStartedPulling="2025-12-01 22:27:23.434543849 +0000 UTC m=+3161.924606166" lastFinishedPulling="2025-12-01 22:27:29.187353957 +0000 UTC m=+3167.677416274" observedRunningTime="2025-12-01 22:27:29.593174181 +0000 UTC m=+3168.083236518" watchObservedRunningTime="2025-12-01 22:27:29.600209795 +0000 UTC m=+3168.090272112" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.630220 4857 scope.go:117] "RemoveContainer" containerID="cb2192707cd1882c643688790bbc8111c00787d69d16d549232bd81f26c9e7d8" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.661964 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.661997 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed82b79a-baa2-42ed-a35c-935dfacc778f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.669724 4857 scope.go:117] "RemoveContainer" containerID="1953a1657ee82da7d20d764e11157274898e04446c6a3dd4976b6233c0e80149" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.732913 4857 scope.go:117] "RemoveContainer" containerID="85615bd37db3ac0824557104e6025e64e6f8570c228d3a62f90bb002e61e6d86" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.901667 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.910018 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.922696 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 22:27:29 crc kubenswrapper[4857]: E1201 22:27:29.923078 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="proxy-httpd" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.923094 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="proxy-httpd" Dec 01 22:27:29 crc kubenswrapper[4857]: E1201 22:27:29.923105 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="sg-core" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.923113 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="sg-core" Dec 01 22:27:29 crc kubenswrapper[4857]: E1201 22:27:29.923138 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="ceilometer-notification-agent" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.923144 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="ceilometer-notification-agent" Dec 01 22:27:29 crc kubenswrapper[4857]: E1201 22:27:29.923163 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="ceilometer-central-agent" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.923168 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="ceilometer-central-agent" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.923349 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="proxy-httpd" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.923362 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="ceilometer-notification-agent" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.923382 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="sg-core" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.923394 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" containerName="ceilometer-central-agent" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.924953 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.927290 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.927625 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.928002 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 22:27:29 crc kubenswrapper[4857]: I1201 22:27:29.938461 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.072349 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-run-httpd\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.072417 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.072523 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-log-httpd\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.072615 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-config-data\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.072666 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.072732 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6cjk\" (UniqueName: \"kubernetes.io/projected/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-kube-api-access-f6cjk\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.072848 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.072910 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-scripts\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.174944 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.175010 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-log-httpd\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.175081 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-config-data\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.175112 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.175171 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6cjk\" (UniqueName: \"kubernetes.io/projected/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-kube-api-access-f6cjk\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.175228 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.175254 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-scripts\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.175314 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-run-httpd\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.175894 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-run-httpd\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.176518 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-log-httpd\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.180745 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.180757 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.181296 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.181437 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-scripts\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.181639 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-config-data\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.196892 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6cjk\" (UniqueName: \"kubernetes.io/projected/ccfb3129-79e9-4397-93c3-05ddbd0e6f76-kube-api-access-f6cjk\") pod \"ceilometer-0\" (UID: \"ccfb3129-79e9-4397-93c3-05ddbd0e6f76\") " pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.241296 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.596144 4857 generic.go:334] "Generic (PLEG): container finished" podID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerID="06ce657f9604cee3e2f96b8b52ced69d57105d8a456835960424448358c665fc" exitCode=0 Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.596429 4857 generic.go:334] "Generic (PLEG): container finished" podID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerID="996fefaac23ac63ff0afd65bea71df4c08d41a78268be04474cc652807258715" exitCode=0 Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.596309 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"eaa5cc4c-c065-45ab-a735-2dfd6af03afe","Type":"ContainerDied","Data":"06ce657f9604cee3e2f96b8b52ced69d57105d8a456835960424448358c665fc"} Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.596491 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"eaa5cc4c-c065-45ab-a735-2dfd6af03afe","Type":"ContainerDied","Data":"996fefaac23ac63ff0afd65bea71df4c08d41a78268be04474cc652807258715"} Dec 01 22:27:30 crc kubenswrapper[4857]: I1201 22:27:30.692035 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 22:27:31 crc kubenswrapper[4857]: I1201 22:27:31.607091 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccfb3129-79e9-4397-93c3-05ddbd0e6f76","Type":"ContainerStarted","Data":"1ad00c82e8cb4a7ed0375b37ab19971a93d4a06421df37fabf0c16fb748bf4da"} Dec 01 22:27:31 crc kubenswrapper[4857]: I1201 22:27:31.850598 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed82b79a-baa2-42ed-a35c-935dfacc778f" path="/var/lib/kubelet/pods/ed82b79a-baa2-42ed-a35c-935dfacc778f/volumes" Dec 01 22:27:32 crc kubenswrapper[4857]: I1201 22:27:32.637646 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccfb3129-79e9-4397-93c3-05ddbd0e6f76","Type":"ContainerStarted","Data":"d5702dc2499e99f87e94ea5c497356f35bdeef779ee027c7522a91bd8eeab98c"} Dec 01 22:27:33 crc kubenswrapper[4857]: I1201 22:27:33.647103 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccfb3129-79e9-4397-93c3-05ddbd0e6f76","Type":"ContainerStarted","Data":"ef8050bd9ff0910bc03634456585068c999c8d08222d93f1a90db425c56d7057"} Dec 01 22:27:34 crc kubenswrapper[4857]: I1201 22:27:34.669667 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccfb3129-79e9-4397-93c3-05ddbd0e6f76","Type":"ContainerStarted","Data":"f1fc118cf4c07cfe92daffbcb10e1b31c0d460ad7be1dcb4693eaf1bc6a4cc4c"} Dec 01 22:27:35 crc kubenswrapper[4857]: I1201 22:27:35.683433 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccfb3129-79e9-4397-93c3-05ddbd0e6f76","Type":"ContainerStarted","Data":"e724d5c5637f0071bfb194c96907426403915be39fc380e600cbf013f24c5c22"} Dec 01 22:27:35 crc kubenswrapper[4857]: I1201 22:27:35.683833 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 22:27:35 crc kubenswrapper[4857]: I1201 22:27:35.709866 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.123446095 podStartE2EDuration="6.709849193s" podCreationTimestamp="2025-12-01 22:27:29 +0000 UTC" firstStartedPulling="2025-12-01 22:27:30.699364668 +0000 UTC m=+3169.189426985" lastFinishedPulling="2025-12-01 22:27:35.285767756 +0000 UTC m=+3173.775830083" observedRunningTime="2025-12-01 22:27:35.708548481 +0000 UTC m=+3174.198610798" watchObservedRunningTime="2025-12-01 22:27:35.709849193 +0000 UTC m=+3174.199911510" Dec 01 22:27:37 crc kubenswrapper[4857]: I1201 22:27:37.836614 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:27:37 crc kubenswrapper[4857]: E1201 22:27:37.839732 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:27:50 crc kubenswrapper[4857]: I1201 22:27:50.911774 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:27:50 crc kubenswrapper[4857]: E1201 22:27:50.912437 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.079226 4857 generic.go:334] "Generic (PLEG): container finished" podID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerID="d39378bf79c21b3adfa86e270c6cd68b9434e065b3e0a66169d33e26935bd602" exitCode=137 Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.079951 4857 generic.go:334] "Generic (PLEG): container finished" podID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerID="4702163fa4c65d550ec70cb4b1c15e4c93137d835ae245ef328119d26963f795" exitCode=137 Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.079984 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"eaa5cc4c-c065-45ab-a735-2dfd6af03afe","Type":"ContainerDied","Data":"d39378bf79c21b3adfa86e270c6cd68b9434e065b3e0a66169d33e26935bd602"} Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.080021 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"eaa5cc4c-c065-45ab-a735-2dfd6af03afe","Type":"ContainerDied","Data":"4702163fa4c65d550ec70cb4b1c15e4c93137d835ae245ef328119d26963f795"} Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.080037 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"eaa5cc4c-c065-45ab-a735-2dfd6af03afe","Type":"ContainerDied","Data":"225850ee9a1a34b0e9b00b3f5c91987cb07de39461dbf1501bb53e5f2aa04bb4"} Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.080076 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="225850ee9a1a34b0e9b00b3f5c91987cb07de39461dbf1501bb53e5f2aa04bb4" Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.129768 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.154848 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb2sz\" (UniqueName: \"kubernetes.io/projected/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-kube-api-access-gb2sz\") pod \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.154990 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-scripts\") pod \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.155226 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-combined-ca-bundle\") pod \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.155449 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-config-data\") pod \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\" (UID: \"eaa5cc4c-c065-45ab-a735-2dfd6af03afe\") " Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.169573 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-scripts" (OuterVolumeSpecName: "scripts") pod "eaa5cc4c-c065-45ab-a735-2dfd6af03afe" (UID: "eaa5cc4c-c065-45ab-a735-2dfd6af03afe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.210840 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-kube-api-access-gb2sz" (OuterVolumeSpecName: "kube-api-access-gb2sz") pod "eaa5cc4c-c065-45ab-a735-2dfd6af03afe" (UID: "eaa5cc4c-c065-45ab-a735-2dfd6af03afe"). InnerVolumeSpecName "kube-api-access-gb2sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.256680 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.262433 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb2sz\" (UniqueName: \"kubernetes.io/projected/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-kube-api-access-gb2sz\") on node \"crc\" DevicePath \"\"" Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.264994 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.327532 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-config-data" (OuterVolumeSpecName: "config-data") pod "eaa5cc4c-c065-45ab-a735-2dfd6af03afe" (UID: "eaa5cc4c-c065-45ab-a735-2dfd6af03afe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.344673 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaa5cc4c-c065-45ab-a735-2dfd6af03afe" (UID: "eaa5cc4c-c065-45ab-a735-2dfd6af03afe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.366752 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:28:00 crc kubenswrapper[4857]: I1201 22:28:00.366786 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa5cc4c-c065-45ab-a735-2dfd6af03afe-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.088143 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.125482 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.141235 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.156494 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 01 22:28:01 crc kubenswrapper[4857]: E1201 22:28:01.156977 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-notifier" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.156996 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-notifier" Dec 01 22:28:01 crc kubenswrapper[4857]: E1201 22:28:01.157025 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-evaluator" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.157032 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-evaluator" Dec 01 22:28:01 crc kubenswrapper[4857]: E1201 22:28:01.157067 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-api" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.157076 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-api" Dec 01 22:28:01 crc kubenswrapper[4857]: E1201 22:28:01.157097 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-listener" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.157102 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-listener" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.157298 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-api" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.157314 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-notifier" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.157326 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-listener" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.157339 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" containerName="aodh-evaluator" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.159108 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.161930 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.162142 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.162453 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-j6fzf" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.162637 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.162800 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.175169 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.307395 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.307737 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-scripts\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.307899 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-internal-tls-certs\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.308008 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-config-data\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.308194 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtgvt\" (UniqueName: \"kubernetes.io/projected/e8c00e5e-1174-4498-8242-3b6c5b765f00-kube-api-access-xtgvt\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.308300 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-public-tls-certs\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.410679 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.411130 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-scripts\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.411240 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-internal-tls-certs\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.411301 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-config-data\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.411420 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtgvt\" (UniqueName: \"kubernetes.io/projected/e8c00e5e-1174-4498-8242-3b6c5b765f00-kube-api-access-xtgvt\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.411477 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-public-tls-certs\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.415447 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.416180 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-scripts\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.416799 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-public-tls-certs\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.418176 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-internal-tls-certs\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.420096 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-config-data\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.432139 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtgvt\" (UniqueName: \"kubernetes.io/projected/e8c00e5e-1174-4498-8242-3b6c5b765f00-kube-api-access-xtgvt\") pod \"aodh-0\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.480521 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.846314 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaa5cc4c-c065-45ab-a735-2dfd6af03afe" path="/var/lib/kubelet/pods/eaa5cc4c-c065-45ab-a735-2dfd6af03afe/volumes" Dec 01 22:28:01 crc kubenswrapper[4857]: I1201 22:28:01.974828 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 22:28:01 crc kubenswrapper[4857]: W1201 22:28:01.978984 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8c00e5e_1174_4498_8242_3b6c5b765f00.slice/crio-404f22886916e56c907db5f2753cbed85d04ed13ff5ee5ac2157fb56c676f8fd WatchSource:0}: Error finding container 404f22886916e56c907db5f2753cbed85d04ed13ff5ee5ac2157fb56c676f8fd: Status 404 returned error can't find the container with id 404f22886916e56c907db5f2753cbed85d04ed13ff5ee5ac2157fb56c676f8fd Dec 01 22:28:02 crc kubenswrapper[4857]: I1201 22:28:02.099825 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e8c00e5e-1174-4498-8242-3b6c5b765f00","Type":"ContainerStarted","Data":"404f22886916e56c907db5f2753cbed85d04ed13ff5ee5ac2157fb56c676f8fd"} Dec 01 22:28:03 crc kubenswrapper[4857]: I1201 22:28:03.110912 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e8c00e5e-1174-4498-8242-3b6c5b765f00","Type":"ContainerStarted","Data":"3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3"} Dec 01 22:28:04 crc kubenswrapper[4857]: I1201 22:28:04.128820 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e8c00e5e-1174-4498-8242-3b6c5b765f00","Type":"ContainerStarted","Data":"641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e"} Dec 01 22:28:04 crc kubenswrapper[4857]: I1201 22:28:04.836417 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:28:04 crc kubenswrapper[4857]: E1201 22:28:04.837165 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:28:05 crc kubenswrapper[4857]: I1201 22:28:05.138466 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e8c00e5e-1174-4498-8242-3b6c5b765f00","Type":"ContainerStarted","Data":"bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc"} Dec 01 22:28:06 crc kubenswrapper[4857]: I1201 22:28:06.151710 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e8c00e5e-1174-4498-8242-3b6c5b765f00","Type":"ContainerStarted","Data":"ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee"} Dec 01 22:28:06 crc kubenswrapper[4857]: I1201 22:28:06.194297 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.7818695820000001 podStartE2EDuration="5.194274085s" podCreationTimestamp="2025-12-01 22:28:01 +0000 UTC" firstStartedPulling="2025-12-01 22:28:01.981893192 +0000 UTC m=+3200.471955519" lastFinishedPulling="2025-12-01 22:28:05.394297705 +0000 UTC m=+3203.884360022" observedRunningTime="2025-12-01 22:28:06.178536085 +0000 UTC m=+3204.668598402" watchObservedRunningTime="2025-12-01 22:28:06.194274085 +0000 UTC m=+3204.684336402" Dec 01 22:28:18 crc kubenswrapper[4857]: I1201 22:28:18.836414 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:28:18 crc kubenswrapper[4857]: E1201 22:28:18.839622 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:28:30 crc kubenswrapper[4857]: I1201 22:28:30.837527 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:28:30 crc kubenswrapper[4857]: E1201 22:28:30.839507 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:28:45 crc kubenswrapper[4857]: I1201 22:28:45.836513 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:28:45 crc kubenswrapper[4857]: E1201 22:28:45.838139 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:28:58 crc kubenswrapper[4857]: I1201 22:28:58.835365 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:28:58 crc kubenswrapper[4857]: E1201 22:28:58.836840 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.499143 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9dmnz"] Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.503923 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.514592 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9dmnz"] Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.596246 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8104112e-baa9-4300-89e8-2e3e35cb520a-utilities\") pod \"redhat-operators-9dmnz\" (UID: \"8104112e-baa9-4300-89e8-2e3e35cb520a\") " pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.596332 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r42vv\" (UniqueName: \"kubernetes.io/projected/8104112e-baa9-4300-89e8-2e3e35cb520a-kube-api-access-r42vv\") pod \"redhat-operators-9dmnz\" (UID: \"8104112e-baa9-4300-89e8-2e3e35cb520a\") " pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.596576 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8104112e-baa9-4300-89e8-2e3e35cb520a-catalog-content\") pod \"redhat-operators-9dmnz\" (UID: \"8104112e-baa9-4300-89e8-2e3e35cb520a\") " pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.698208 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8104112e-baa9-4300-89e8-2e3e35cb520a-utilities\") pod \"redhat-operators-9dmnz\" (UID: \"8104112e-baa9-4300-89e8-2e3e35cb520a\") " pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.698282 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r42vv\" (UniqueName: \"kubernetes.io/projected/8104112e-baa9-4300-89e8-2e3e35cb520a-kube-api-access-r42vv\") pod \"redhat-operators-9dmnz\" (UID: \"8104112e-baa9-4300-89e8-2e3e35cb520a\") " pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.698347 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8104112e-baa9-4300-89e8-2e3e35cb520a-catalog-content\") pod \"redhat-operators-9dmnz\" (UID: \"8104112e-baa9-4300-89e8-2e3e35cb520a\") " pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.698841 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8104112e-baa9-4300-89e8-2e3e35cb520a-catalog-content\") pod \"redhat-operators-9dmnz\" (UID: \"8104112e-baa9-4300-89e8-2e3e35cb520a\") " pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.699101 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8104112e-baa9-4300-89e8-2e3e35cb520a-utilities\") pod \"redhat-operators-9dmnz\" (UID: \"8104112e-baa9-4300-89e8-2e3e35cb520a\") " pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.728379 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r42vv\" (UniqueName: \"kubernetes.io/projected/8104112e-baa9-4300-89e8-2e3e35cb520a-kube-api-access-r42vv\") pod \"redhat-operators-9dmnz\" (UID: \"8104112e-baa9-4300-89e8-2e3e35cb520a\") " pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:04 crc kubenswrapper[4857]: I1201 22:29:04.845883 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:05 crc kubenswrapper[4857]: I1201 22:29:05.120849 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9dmnz"] Dec 01 22:29:05 crc kubenswrapper[4857]: I1201 22:29:05.221103 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dmnz" event={"ID":"8104112e-baa9-4300-89e8-2e3e35cb520a","Type":"ContainerStarted","Data":"ca5f9a66144bbc7426dac650251efe8d29ad4d3c04d3776759896f027e4b9711"} Dec 01 22:29:06 crc kubenswrapper[4857]: I1201 22:29:06.246129 4857 generic.go:334] "Generic (PLEG): container finished" podID="8104112e-baa9-4300-89e8-2e3e35cb520a" containerID="5fe7a8278b57443deb550ef862aed1153f4c6241e407ae30a4febc83cbc4426f" exitCode=0 Dec 01 22:29:06 crc kubenswrapper[4857]: I1201 22:29:06.246378 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dmnz" event={"ID":"8104112e-baa9-4300-89e8-2e3e35cb520a","Type":"ContainerDied","Data":"5fe7a8278b57443deb550ef862aed1153f4c6241e407ae30a4febc83cbc4426f"} Dec 01 22:29:06 crc kubenswrapper[4857]: I1201 22:29:06.249451 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 22:29:08 crc kubenswrapper[4857]: I1201 22:29:08.272269 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dmnz" event={"ID":"8104112e-baa9-4300-89e8-2e3e35cb520a","Type":"ContainerStarted","Data":"fc7c7b375c99886a563908f291b42a9d667b5add8ded0b8c8b9060eb99f7694f"} Dec 01 22:29:10 crc kubenswrapper[4857]: I1201 22:29:10.299346 4857 generic.go:334] "Generic (PLEG): container finished" podID="8104112e-baa9-4300-89e8-2e3e35cb520a" containerID="fc7c7b375c99886a563908f291b42a9d667b5add8ded0b8c8b9060eb99f7694f" exitCode=0 Dec 01 22:29:10 crc kubenswrapper[4857]: I1201 22:29:10.299445 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dmnz" event={"ID":"8104112e-baa9-4300-89e8-2e3e35cb520a","Type":"ContainerDied","Data":"fc7c7b375c99886a563908f291b42a9d667b5add8ded0b8c8b9060eb99f7694f"} Dec 01 22:29:11 crc kubenswrapper[4857]: I1201 22:29:11.311660 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dmnz" event={"ID":"8104112e-baa9-4300-89e8-2e3e35cb520a","Type":"ContainerStarted","Data":"a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad"} Dec 01 22:29:11 crc kubenswrapper[4857]: I1201 22:29:11.351731 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9dmnz" podStartSLOduration=2.671028266 podStartE2EDuration="7.351709322s" podCreationTimestamp="2025-12-01 22:29:04 +0000 UTC" firstStartedPulling="2025-12-01 22:29:06.249162034 +0000 UTC m=+3264.739224361" lastFinishedPulling="2025-12-01 22:29:10.92984306 +0000 UTC m=+3269.419905417" observedRunningTime="2025-12-01 22:29:11.343027507 +0000 UTC m=+3269.833089844" watchObservedRunningTime="2025-12-01 22:29:11.351709322 +0000 UTC m=+3269.841771759" Dec 01 22:29:11 crc kubenswrapper[4857]: I1201 22:29:11.845678 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:29:11 crc kubenswrapper[4857]: E1201 22:29:11.846137 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:29:14 crc kubenswrapper[4857]: I1201 22:29:14.846099 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:14 crc kubenswrapper[4857]: I1201 22:29:14.846387 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:15 crc kubenswrapper[4857]: I1201 22:29:15.922366 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9dmnz" podUID="8104112e-baa9-4300-89e8-2e3e35cb520a" containerName="registry-server" probeResult="failure" output=< Dec 01 22:29:15 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Dec 01 22:29:15 crc kubenswrapper[4857]: > Dec 01 22:29:23 crc kubenswrapper[4857]: I1201 22:29:23.184676 4857 scope.go:117] "RemoveContainer" containerID="bcc09cc92b460770758f3386ee47427332090208240c88b4ae69c369744c60ca" Dec 01 22:29:24 crc kubenswrapper[4857]: I1201 22:29:24.919161 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:24 crc kubenswrapper[4857]: I1201 22:29:24.992251 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:25 crc kubenswrapper[4857]: I1201 22:29:25.175838 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9dmnz"] Dec 01 22:29:25 crc kubenswrapper[4857]: I1201 22:29:25.850080 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:29:25 crc kubenswrapper[4857]: E1201 22:29:25.850356 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:29:26 crc kubenswrapper[4857]: I1201 22:29:26.484463 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9dmnz" podUID="8104112e-baa9-4300-89e8-2e3e35cb520a" containerName="registry-server" containerID="cri-o://a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad" gracePeriod=2 Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.053100 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.078795 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8104112e-baa9-4300-89e8-2e3e35cb520a-catalog-content\") pod \"8104112e-baa9-4300-89e8-2e3e35cb520a\" (UID: \"8104112e-baa9-4300-89e8-2e3e35cb520a\") " Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.078991 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8104112e-baa9-4300-89e8-2e3e35cb520a-utilities\") pod \"8104112e-baa9-4300-89e8-2e3e35cb520a\" (UID: \"8104112e-baa9-4300-89e8-2e3e35cb520a\") " Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.079090 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r42vv\" (UniqueName: \"kubernetes.io/projected/8104112e-baa9-4300-89e8-2e3e35cb520a-kube-api-access-r42vv\") pod \"8104112e-baa9-4300-89e8-2e3e35cb520a\" (UID: \"8104112e-baa9-4300-89e8-2e3e35cb520a\") " Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.079928 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8104112e-baa9-4300-89e8-2e3e35cb520a-utilities" (OuterVolumeSpecName: "utilities") pod "8104112e-baa9-4300-89e8-2e3e35cb520a" (UID: "8104112e-baa9-4300-89e8-2e3e35cb520a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.097381 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8104112e-baa9-4300-89e8-2e3e35cb520a-kube-api-access-r42vv" (OuterVolumeSpecName: "kube-api-access-r42vv") pod "8104112e-baa9-4300-89e8-2e3e35cb520a" (UID: "8104112e-baa9-4300-89e8-2e3e35cb520a"). InnerVolumeSpecName "kube-api-access-r42vv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.177403 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8104112e-baa9-4300-89e8-2e3e35cb520a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8104112e-baa9-4300-89e8-2e3e35cb520a" (UID: "8104112e-baa9-4300-89e8-2e3e35cb520a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.181421 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8104112e-baa9-4300-89e8-2e3e35cb520a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.181453 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8104112e-baa9-4300-89e8-2e3e35cb520a-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.181464 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r42vv\" (UniqueName: \"kubernetes.io/projected/8104112e-baa9-4300-89e8-2e3e35cb520a-kube-api-access-r42vv\") on node \"crc\" DevicePath \"\"" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.502947 4857 generic.go:334] "Generic (PLEG): container finished" podID="8104112e-baa9-4300-89e8-2e3e35cb520a" containerID="a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad" exitCode=0 Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.502998 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dmnz" event={"ID":"8104112e-baa9-4300-89e8-2e3e35cb520a","Type":"ContainerDied","Data":"a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad"} Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.503033 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dmnz" event={"ID":"8104112e-baa9-4300-89e8-2e3e35cb520a","Type":"ContainerDied","Data":"ca5f9a66144bbc7426dac650251efe8d29ad4d3c04d3776759896f027e4b9711"} Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.503077 4857 scope.go:117] "RemoveContainer" containerID="a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.503260 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dmnz" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.544730 4857 scope.go:117] "RemoveContainer" containerID="fc7c7b375c99886a563908f291b42a9d667b5add8ded0b8c8b9060eb99f7694f" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.562376 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9dmnz"] Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.575721 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9dmnz"] Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.583078 4857 scope.go:117] "RemoveContainer" containerID="5fe7a8278b57443deb550ef862aed1153f4c6241e407ae30a4febc83cbc4426f" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.660953 4857 scope.go:117] "RemoveContainer" containerID="a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad" Dec 01 22:29:27 crc kubenswrapper[4857]: E1201 22:29:27.661665 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad\": container with ID starting with a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad not found: ID does not exist" containerID="a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.661724 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad"} err="failed to get container status \"a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad\": rpc error: code = NotFound desc = could not find container \"a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad\": container with ID starting with a953f9c754cb146478a03dfefdfe2c2b277ef71c777d065e7b4365fdc01ba4ad not found: ID does not exist" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.661766 4857 scope.go:117] "RemoveContainer" containerID="fc7c7b375c99886a563908f291b42a9d667b5add8ded0b8c8b9060eb99f7694f" Dec 01 22:29:27 crc kubenswrapper[4857]: E1201 22:29:27.662221 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc7c7b375c99886a563908f291b42a9d667b5add8ded0b8c8b9060eb99f7694f\": container with ID starting with fc7c7b375c99886a563908f291b42a9d667b5add8ded0b8c8b9060eb99f7694f not found: ID does not exist" containerID="fc7c7b375c99886a563908f291b42a9d667b5add8ded0b8c8b9060eb99f7694f" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.662265 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc7c7b375c99886a563908f291b42a9d667b5add8ded0b8c8b9060eb99f7694f"} err="failed to get container status \"fc7c7b375c99886a563908f291b42a9d667b5add8ded0b8c8b9060eb99f7694f\": rpc error: code = NotFound desc = could not find container \"fc7c7b375c99886a563908f291b42a9d667b5add8ded0b8c8b9060eb99f7694f\": container with ID starting with fc7c7b375c99886a563908f291b42a9d667b5add8ded0b8c8b9060eb99f7694f not found: ID does not exist" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.662296 4857 scope.go:117] "RemoveContainer" containerID="5fe7a8278b57443deb550ef862aed1153f4c6241e407ae30a4febc83cbc4426f" Dec 01 22:29:27 crc kubenswrapper[4857]: E1201 22:29:27.662714 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fe7a8278b57443deb550ef862aed1153f4c6241e407ae30a4febc83cbc4426f\": container with ID starting with 5fe7a8278b57443deb550ef862aed1153f4c6241e407ae30a4febc83cbc4426f not found: ID does not exist" containerID="5fe7a8278b57443deb550ef862aed1153f4c6241e407ae30a4febc83cbc4426f" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.662754 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fe7a8278b57443deb550ef862aed1153f4c6241e407ae30a4febc83cbc4426f"} err="failed to get container status \"5fe7a8278b57443deb550ef862aed1153f4c6241e407ae30a4febc83cbc4426f\": rpc error: code = NotFound desc = could not find container \"5fe7a8278b57443deb550ef862aed1153f4c6241e407ae30a4febc83cbc4426f\": container with ID starting with 5fe7a8278b57443deb550ef862aed1153f4c6241e407ae30a4febc83cbc4426f not found: ID does not exist" Dec 01 22:29:27 crc kubenswrapper[4857]: I1201 22:29:27.854570 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8104112e-baa9-4300-89e8-2e3e35cb520a" path="/var/lib/kubelet/pods/8104112e-baa9-4300-89e8-2e3e35cb520a/volumes" Dec 01 22:29:37 crc kubenswrapper[4857]: I1201 22:29:37.835455 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:29:37 crc kubenswrapper[4857]: E1201 22:29:37.837358 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:29:50 crc kubenswrapper[4857]: I1201 22:29:50.835883 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:29:50 crc kubenswrapper[4857]: E1201 22:29:50.836757 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.176381 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h"] Dec 01 22:30:00 crc kubenswrapper[4857]: E1201 22:30:00.177809 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8104112e-baa9-4300-89e8-2e3e35cb520a" containerName="extract-utilities" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.177837 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8104112e-baa9-4300-89e8-2e3e35cb520a" containerName="extract-utilities" Dec 01 22:30:00 crc kubenswrapper[4857]: E1201 22:30:00.177865 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8104112e-baa9-4300-89e8-2e3e35cb520a" containerName="registry-server" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.177877 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8104112e-baa9-4300-89e8-2e3e35cb520a" containerName="registry-server" Dec 01 22:30:00 crc kubenswrapper[4857]: E1201 22:30:00.177908 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8104112e-baa9-4300-89e8-2e3e35cb520a" containerName="extract-content" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.177919 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8104112e-baa9-4300-89e8-2e3e35cb520a" containerName="extract-content" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.178297 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="8104112e-baa9-4300-89e8-2e3e35cb520a" containerName="registry-server" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.179467 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.183451 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.183943 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.214138 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h"] Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.270123 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed4ab2af-857a-4db2-b31e-af89b15e7f49-secret-volume\") pod \"collect-profiles-29410470-ssr5h\" (UID: \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.270291 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed4ab2af-857a-4db2-b31e-af89b15e7f49-config-volume\") pod \"collect-profiles-29410470-ssr5h\" (UID: \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.270683 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khk96\" (UniqueName: \"kubernetes.io/projected/ed4ab2af-857a-4db2-b31e-af89b15e7f49-kube-api-access-khk96\") pod \"collect-profiles-29410470-ssr5h\" (UID: \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.373032 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed4ab2af-857a-4db2-b31e-af89b15e7f49-secret-volume\") pod \"collect-profiles-29410470-ssr5h\" (UID: \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.373502 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed4ab2af-857a-4db2-b31e-af89b15e7f49-config-volume\") pod \"collect-profiles-29410470-ssr5h\" (UID: \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.373701 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khk96\" (UniqueName: \"kubernetes.io/projected/ed4ab2af-857a-4db2-b31e-af89b15e7f49-kube-api-access-khk96\") pod \"collect-profiles-29410470-ssr5h\" (UID: \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.376286 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed4ab2af-857a-4db2-b31e-af89b15e7f49-config-volume\") pod \"collect-profiles-29410470-ssr5h\" (UID: \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.387486 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed4ab2af-857a-4db2-b31e-af89b15e7f49-secret-volume\") pod \"collect-profiles-29410470-ssr5h\" (UID: \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.409464 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khk96\" (UniqueName: \"kubernetes.io/projected/ed4ab2af-857a-4db2-b31e-af89b15e7f49-kube-api-access-khk96\") pod \"collect-profiles-29410470-ssr5h\" (UID: \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:00 crc kubenswrapper[4857]: I1201 22:30:00.512720 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:01 crc kubenswrapper[4857]: I1201 22:30:01.097102 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h"] Dec 01 22:30:01 crc kubenswrapper[4857]: I1201 22:30:01.901118 4857 generic.go:334] "Generic (PLEG): container finished" podID="ed4ab2af-857a-4db2-b31e-af89b15e7f49" containerID="060e2455c17a2eea114694b0b89bafe3d14b57677b660b7fa8fab8665f4e719c" exitCode=0 Dec 01 22:30:01 crc kubenswrapper[4857]: I1201 22:30:01.901245 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" event={"ID":"ed4ab2af-857a-4db2-b31e-af89b15e7f49","Type":"ContainerDied","Data":"060e2455c17a2eea114694b0b89bafe3d14b57677b660b7fa8fab8665f4e719c"} Dec 01 22:30:01 crc kubenswrapper[4857]: I1201 22:30:01.901592 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" event={"ID":"ed4ab2af-857a-4db2-b31e-af89b15e7f49","Type":"ContainerStarted","Data":"f061e76c53f520fef2e997315a1ef43d1faef48bf5ec4b272d6b811ef88000d7"} Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.280796 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.345275 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khk96\" (UniqueName: \"kubernetes.io/projected/ed4ab2af-857a-4db2-b31e-af89b15e7f49-kube-api-access-khk96\") pod \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\" (UID: \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\") " Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.345447 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed4ab2af-857a-4db2-b31e-af89b15e7f49-secret-volume\") pod \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\" (UID: \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\") " Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.345501 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed4ab2af-857a-4db2-b31e-af89b15e7f49-config-volume\") pod \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\" (UID: \"ed4ab2af-857a-4db2-b31e-af89b15e7f49\") " Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.346686 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed4ab2af-857a-4db2-b31e-af89b15e7f49-config-volume" (OuterVolumeSpecName: "config-volume") pod "ed4ab2af-857a-4db2-b31e-af89b15e7f49" (UID: "ed4ab2af-857a-4db2-b31e-af89b15e7f49"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.353607 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed4ab2af-857a-4db2-b31e-af89b15e7f49-kube-api-access-khk96" (OuterVolumeSpecName: "kube-api-access-khk96") pod "ed4ab2af-857a-4db2-b31e-af89b15e7f49" (UID: "ed4ab2af-857a-4db2-b31e-af89b15e7f49"). InnerVolumeSpecName "kube-api-access-khk96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.353700 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed4ab2af-857a-4db2-b31e-af89b15e7f49-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ed4ab2af-857a-4db2-b31e-af89b15e7f49" (UID: "ed4ab2af-857a-4db2-b31e-af89b15e7f49"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.448020 4857 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed4ab2af-857a-4db2-b31e-af89b15e7f49-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.448096 4857 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed4ab2af-857a-4db2-b31e-af89b15e7f49-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.448114 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khk96\" (UniqueName: \"kubernetes.io/projected/ed4ab2af-857a-4db2-b31e-af89b15e7f49-kube-api-access-khk96\") on node \"crc\" DevicePath \"\"" Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.837192 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:30:03 crc kubenswrapper[4857]: E1201 22:30:03.838200 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.937345 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" event={"ID":"ed4ab2af-857a-4db2-b31e-af89b15e7f49","Type":"ContainerDied","Data":"f061e76c53f520fef2e997315a1ef43d1faef48bf5ec4b272d6b811ef88000d7"} Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.937405 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f061e76c53f520fef2e997315a1ef43d1faef48bf5ec4b272d6b811ef88000d7" Dec 01 22:30:03 crc kubenswrapper[4857]: I1201 22:30:03.937445 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410470-ssr5h" Dec 01 22:30:04 crc kubenswrapper[4857]: I1201 22:30:04.413861 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27"] Dec 01 22:30:04 crc kubenswrapper[4857]: I1201 22:30:04.424909 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410425-q7l27"] Dec 01 22:30:05 crc kubenswrapper[4857]: I1201 22:30:05.848840 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76127279-fe59-49e5-82cb-8f2d2a8afbe2" path="/var/lib/kubelet/pods/76127279-fe59-49e5-82cb-8f2d2a8afbe2/volumes" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.193303 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xhhdw"] Dec 01 22:30:14 crc kubenswrapper[4857]: E1201 22:30:14.194672 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed4ab2af-857a-4db2-b31e-af89b15e7f49" containerName="collect-profiles" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.194703 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed4ab2af-857a-4db2-b31e-af89b15e7f49" containerName="collect-profiles" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.195481 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed4ab2af-857a-4db2-b31e-af89b15e7f49" containerName="collect-profiles" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.198388 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.226948 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhhdw"] Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.333395 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-utilities\") pod \"redhat-marketplace-xhhdw\" (UID: \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\") " pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.333453 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bptw4\" (UniqueName: \"kubernetes.io/projected/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-kube-api-access-bptw4\") pod \"redhat-marketplace-xhhdw\" (UID: \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\") " pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.333748 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-catalog-content\") pod \"redhat-marketplace-xhhdw\" (UID: \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\") " pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.436309 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-utilities\") pod \"redhat-marketplace-xhhdw\" (UID: \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\") " pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.436363 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bptw4\" (UniqueName: \"kubernetes.io/projected/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-kube-api-access-bptw4\") pod \"redhat-marketplace-xhhdw\" (UID: \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\") " pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.436610 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-catalog-content\") pod \"redhat-marketplace-xhhdw\" (UID: \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\") " pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.436890 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-utilities\") pod \"redhat-marketplace-xhhdw\" (UID: \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\") " pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.437150 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-catalog-content\") pod \"redhat-marketplace-xhhdw\" (UID: \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\") " pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.469895 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bptw4\" (UniqueName: \"kubernetes.io/projected/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-kube-api-access-bptw4\") pod \"redhat-marketplace-xhhdw\" (UID: \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\") " pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:14 crc kubenswrapper[4857]: I1201 22:30:14.538123 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:15 crc kubenswrapper[4857]: W1201 22:30:15.086816 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80e32b78_a7f9_4cfb_9b38_80fd18405b6e.slice/crio-090be4081efeba508c8e5f3c6e751cbdc88ce8914c5ca2537e4c3362aa47a495 WatchSource:0}: Error finding container 090be4081efeba508c8e5f3c6e751cbdc88ce8914c5ca2537e4c3362aa47a495: Status 404 returned error can't find the container with id 090be4081efeba508c8e5f3c6e751cbdc88ce8914c5ca2537e4c3362aa47a495 Dec 01 22:30:15 crc kubenswrapper[4857]: I1201 22:30:15.088285 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhhdw"] Dec 01 22:30:16 crc kubenswrapper[4857]: I1201 22:30:16.072497 4857 generic.go:334] "Generic (PLEG): container finished" podID="80e32b78-a7f9-4cfb-9b38-80fd18405b6e" containerID="20c90a9a9e62b28b62e500818a098fa0822272995a7abb922282b26a9379bebe" exitCode=0 Dec 01 22:30:16 crc kubenswrapper[4857]: I1201 22:30:16.073097 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhhdw" event={"ID":"80e32b78-a7f9-4cfb-9b38-80fd18405b6e","Type":"ContainerDied","Data":"20c90a9a9e62b28b62e500818a098fa0822272995a7abb922282b26a9379bebe"} Dec 01 22:30:16 crc kubenswrapper[4857]: I1201 22:30:16.073157 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhhdw" event={"ID":"80e32b78-a7f9-4cfb-9b38-80fd18405b6e","Type":"ContainerStarted","Data":"090be4081efeba508c8e5f3c6e751cbdc88ce8914c5ca2537e4c3362aa47a495"} Dec 01 22:30:18 crc kubenswrapper[4857]: I1201 22:30:18.098723 4857 generic.go:334] "Generic (PLEG): container finished" podID="80e32b78-a7f9-4cfb-9b38-80fd18405b6e" containerID="69fa118b133969b20e1b15e33d671eaf1399e5084dcc9696715fcf8b664854dc" exitCode=0 Dec 01 22:30:18 crc kubenswrapper[4857]: I1201 22:30:18.098820 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhhdw" event={"ID":"80e32b78-a7f9-4cfb-9b38-80fd18405b6e","Type":"ContainerDied","Data":"69fa118b133969b20e1b15e33d671eaf1399e5084dcc9696715fcf8b664854dc"} Dec 01 22:30:18 crc kubenswrapper[4857]: I1201 22:30:18.860892 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:30:18 crc kubenswrapper[4857]: E1201 22:30:18.862389 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:30:19 crc kubenswrapper[4857]: I1201 22:30:19.114395 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhhdw" event={"ID":"80e32b78-a7f9-4cfb-9b38-80fd18405b6e","Type":"ContainerStarted","Data":"8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767"} Dec 01 22:30:19 crc kubenswrapper[4857]: I1201 22:30:19.141262 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xhhdw" podStartSLOduration=2.628774898 podStartE2EDuration="5.141242375s" podCreationTimestamp="2025-12-01 22:30:14 +0000 UTC" firstStartedPulling="2025-12-01 22:30:16.075067577 +0000 UTC m=+3334.565129904" lastFinishedPulling="2025-12-01 22:30:18.587535014 +0000 UTC m=+3337.077597381" observedRunningTime="2025-12-01 22:30:19.135703979 +0000 UTC m=+3337.625766336" watchObservedRunningTime="2025-12-01 22:30:19.141242375 +0000 UTC m=+3337.631304692" Dec 01 22:30:23 crc kubenswrapper[4857]: I1201 22:30:23.284125 4857 scope.go:117] "RemoveContainer" containerID="3ad7b82ddc864cae02daef847d2c379ecfcb6459c26da987117dbb1ee24fab69" Dec 01 22:30:23 crc kubenswrapper[4857]: I1201 22:30:23.333272 4857 scope.go:117] "RemoveContainer" containerID="722642a2b92ce978766499c63e078bc163e38c7e3cbb6e4e4647ebcf2973ddfb" Dec 01 22:30:23 crc kubenswrapper[4857]: I1201 22:30:23.387586 4857 scope.go:117] "RemoveContainer" containerID="99fa09734d6859dea0b7d5628e7c2b1a1d0de6adc00243a6343971e35cba79f8" Dec 01 22:30:24 crc kubenswrapper[4857]: I1201 22:30:24.538819 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:24 crc kubenswrapper[4857]: I1201 22:30:24.539124 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:24 crc kubenswrapper[4857]: I1201 22:30:24.620217 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:25 crc kubenswrapper[4857]: I1201 22:30:25.272365 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:25 crc kubenswrapper[4857]: I1201 22:30:25.347936 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhhdw"] Dec 01 22:30:27 crc kubenswrapper[4857]: I1201 22:30:27.216479 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xhhdw" podUID="80e32b78-a7f9-4cfb-9b38-80fd18405b6e" containerName="registry-server" containerID="cri-o://8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767" gracePeriod=2 Dec 01 22:30:27 crc kubenswrapper[4857]: I1201 22:30:27.805422 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:27 crc kubenswrapper[4857]: I1201 22:30:27.977279 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-utilities\") pod \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\" (UID: \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\") " Dec 01 22:30:27 crc kubenswrapper[4857]: I1201 22:30:27.977637 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bptw4\" (UniqueName: \"kubernetes.io/projected/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-kube-api-access-bptw4\") pod \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\" (UID: \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\") " Dec 01 22:30:27 crc kubenswrapper[4857]: I1201 22:30:27.977809 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-catalog-content\") pod \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\" (UID: \"80e32b78-a7f9-4cfb-9b38-80fd18405b6e\") " Dec 01 22:30:27 crc kubenswrapper[4857]: I1201 22:30:27.978987 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-utilities" (OuterVolumeSpecName: "utilities") pod "80e32b78-a7f9-4cfb-9b38-80fd18405b6e" (UID: "80e32b78-a7f9-4cfb-9b38-80fd18405b6e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:30:27 crc kubenswrapper[4857]: I1201 22:30:27.979284 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:30:27 crc kubenswrapper[4857]: I1201 22:30:27.983342 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-kube-api-access-bptw4" (OuterVolumeSpecName: "kube-api-access-bptw4") pod "80e32b78-a7f9-4cfb-9b38-80fd18405b6e" (UID: "80e32b78-a7f9-4cfb-9b38-80fd18405b6e"). InnerVolumeSpecName "kube-api-access-bptw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:30:27 crc kubenswrapper[4857]: I1201 22:30:27.995548 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80e32b78-a7f9-4cfb-9b38-80fd18405b6e" (UID: "80e32b78-a7f9-4cfb-9b38-80fd18405b6e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.081770 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bptw4\" (UniqueName: \"kubernetes.io/projected/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-kube-api-access-bptw4\") on node \"crc\" DevicePath \"\"" Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.081824 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80e32b78-a7f9-4cfb-9b38-80fd18405b6e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.231364 4857 generic.go:334] "Generic (PLEG): container finished" podID="80e32b78-a7f9-4cfb-9b38-80fd18405b6e" containerID="8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767" exitCode=0 Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.231432 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhhdw" event={"ID":"80e32b78-a7f9-4cfb-9b38-80fd18405b6e","Type":"ContainerDied","Data":"8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767"} Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.231464 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhhdw" Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.231482 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhhdw" event={"ID":"80e32b78-a7f9-4cfb-9b38-80fd18405b6e","Type":"ContainerDied","Data":"090be4081efeba508c8e5f3c6e751cbdc88ce8914c5ca2537e4c3362aa47a495"} Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.231515 4857 scope.go:117] "RemoveContainer" containerID="8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767" Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.279667 4857 scope.go:117] "RemoveContainer" containerID="69fa118b133969b20e1b15e33d671eaf1399e5084dcc9696715fcf8b664854dc" Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.292360 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhhdw"] Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.310797 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhhdw"] Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.316633 4857 scope.go:117] "RemoveContainer" containerID="20c90a9a9e62b28b62e500818a098fa0822272995a7abb922282b26a9379bebe" Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.374218 4857 scope.go:117] "RemoveContainer" containerID="8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767" Dec 01 22:30:28 crc kubenswrapper[4857]: E1201 22:30:28.375269 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767\": container with ID starting with 8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767 not found: ID does not exist" containerID="8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767" Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.375618 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767"} err="failed to get container status \"8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767\": rpc error: code = NotFound desc = could not find container \"8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767\": container with ID starting with 8523d824be1da77aa20feeceac51be54e77f577facc67d78d1e804b0e0ebf767 not found: ID does not exist" Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.375798 4857 scope.go:117] "RemoveContainer" containerID="69fa118b133969b20e1b15e33d671eaf1399e5084dcc9696715fcf8b664854dc" Dec 01 22:30:28 crc kubenswrapper[4857]: E1201 22:30:28.376620 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69fa118b133969b20e1b15e33d671eaf1399e5084dcc9696715fcf8b664854dc\": container with ID starting with 69fa118b133969b20e1b15e33d671eaf1399e5084dcc9696715fcf8b664854dc not found: ID does not exist" containerID="69fa118b133969b20e1b15e33d671eaf1399e5084dcc9696715fcf8b664854dc" Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.376684 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69fa118b133969b20e1b15e33d671eaf1399e5084dcc9696715fcf8b664854dc"} err="failed to get container status \"69fa118b133969b20e1b15e33d671eaf1399e5084dcc9696715fcf8b664854dc\": rpc error: code = NotFound desc = could not find container \"69fa118b133969b20e1b15e33d671eaf1399e5084dcc9696715fcf8b664854dc\": container with ID starting with 69fa118b133969b20e1b15e33d671eaf1399e5084dcc9696715fcf8b664854dc not found: ID does not exist" Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.376729 4857 scope.go:117] "RemoveContainer" containerID="20c90a9a9e62b28b62e500818a098fa0822272995a7abb922282b26a9379bebe" Dec 01 22:30:28 crc kubenswrapper[4857]: E1201 22:30:28.377247 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20c90a9a9e62b28b62e500818a098fa0822272995a7abb922282b26a9379bebe\": container with ID starting with 20c90a9a9e62b28b62e500818a098fa0822272995a7abb922282b26a9379bebe not found: ID does not exist" containerID="20c90a9a9e62b28b62e500818a098fa0822272995a7abb922282b26a9379bebe" Dec 01 22:30:28 crc kubenswrapper[4857]: I1201 22:30:28.377294 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20c90a9a9e62b28b62e500818a098fa0822272995a7abb922282b26a9379bebe"} err="failed to get container status \"20c90a9a9e62b28b62e500818a098fa0822272995a7abb922282b26a9379bebe\": rpc error: code = NotFound desc = could not find container \"20c90a9a9e62b28b62e500818a098fa0822272995a7abb922282b26a9379bebe\": container with ID starting with 20c90a9a9e62b28b62e500818a098fa0822272995a7abb922282b26a9379bebe not found: ID does not exist" Dec 01 22:30:29 crc kubenswrapper[4857]: I1201 22:30:29.851576 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80e32b78-a7f9-4cfb-9b38-80fd18405b6e" path="/var/lib/kubelet/pods/80e32b78-a7f9-4cfb-9b38-80fd18405b6e/volumes" Dec 01 22:30:30 crc kubenswrapper[4857]: I1201 22:30:30.835995 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:30:30 crc kubenswrapper[4857]: E1201 22:30:30.836947 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:30:44 crc kubenswrapper[4857]: I1201 22:30:44.836377 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:30:44 crc kubenswrapper[4857]: E1201 22:30:44.837520 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:30:59 crc kubenswrapper[4857]: I1201 22:30:59.836412 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:30:59 crc kubenswrapper[4857]: E1201 22:30:59.837840 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:31:02 crc kubenswrapper[4857]: I1201 22:31:02.797958 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f765fc7d6-x797d_ff230e58-1711-4638-89d6-daf91fb3bfc1/manager/0.log" Dec 01 22:31:13 crc kubenswrapper[4857]: I1201 22:31:13.834790 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:31:13 crc kubenswrapper[4857]: E1201 22:31:13.835628 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:31:16 crc kubenswrapper[4857]: I1201 22:31:16.759598 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd"] Dec 01 22:31:16 crc kubenswrapper[4857]: E1201 22:31:16.760910 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e32b78-a7f9-4cfb-9b38-80fd18405b6e" containerName="extract-utilities" Dec 01 22:31:16 crc kubenswrapper[4857]: I1201 22:31:16.760936 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e32b78-a7f9-4cfb-9b38-80fd18405b6e" containerName="extract-utilities" Dec 01 22:31:16 crc kubenswrapper[4857]: E1201 22:31:16.760999 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e32b78-a7f9-4cfb-9b38-80fd18405b6e" containerName="extract-content" Dec 01 22:31:16 crc kubenswrapper[4857]: I1201 22:31:16.761013 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e32b78-a7f9-4cfb-9b38-80fd18405b6e" containerName="extract-content" Dec 01 22:31:16 crc kubenswrapper[4857]: E1201 22:31:16.761092 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e32b78-a7f9-4cfb-9b38-80fd18405b6e" containerName="registry-server" Dec 01 22:31:16 crc kubenswrapper[4857]: I1201 22:31:16.761108 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e32b78-a7f9-4cfb-9b38-80fd18405b6e" containerName="registry-server" Dec 01 22:31:16 crc kubenswrapper[4857]: I1201 22:31:16.761510 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="80e32b78-a7f9-4cfb-9b38-80fd18405b6e" containerName="registry-server" Dec 01 22:31:16 crc kubenswrapper[4857]: I1201 22:31:16.765305 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:16 crc kubenswrapper[4857]: I1201 22:31:16.769522 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 22:31:16 crc kubenswrapper[4857]: I1201 22:31:16.772429 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd"] Dec 01 22:31:16 crc kubenswrapper[4857]: I1201 22:31:16.938211 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd\" (UID: \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:16 crc kubenswrapper[4857]: I1201 22:31:16.938710 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dnsv\" (UniqueName: \"kubernetes.io/projected/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-kube-api-access-9dnsv\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd\" (UID: \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:16 crc kubenswrapper[4857]: I1201 22:31:16.938839 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd\" (UID: \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:17 crc kubenswrapper[4857]: I1201 22:31:17.040815 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd\" (UID: \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:17 crc kubenswrapper[4857]: I1201 22:31:17.040997 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd\" (UID: \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:17 crc kubenswrapper[4857]: I1201 22:31:17.041263 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dnsv\" (UniqueName: \"kubernetes.io/projected/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-kube-api-access-9dnsv\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd\" (UID: \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:17 crc kubenswrapper[4857]: I1201 22:31:17.041868 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd\" (UID: \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:17 crc kubenswrapper[4857]: I1201 22:31:17.041910 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd\" (UID: \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:17 crc kubenswrapper[4857]: I1201 22:31:17.076499 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dnsv\" (UniqueName: \"kubernetes.io/projected/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-kube-api-access-9dnsv\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd\" (UID: \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:17 crc kubenswrapper[4857]: I1201 22:31:17.090254 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:17 crc kubenswrapper[4857]: I1201 22:31:17.397796 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd"] Dec 01 22:31:17 crc kubenswrapper[4857]: W1201 22:31:17.399279 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda058ee2b_6a45_4dbe_b0dd_1c6c5a0d8ba8.slice/crio-f812681677551a135d5996e9796404664ee0fada500992f9fa3c139a0cf341ea WatchSource:0}: Error finding container f812681677551a135d5996e9796404664ee0fada500992f9fa3c139a0cf341ea: Status 404 returned error can't find the container with id f812681677551a135d5996e9796404664ee0fada500992f9fa3c139a0cf341ea Dec 01 22:31:17 crc kubenswrapper[4857]: I1201 22:31:17.898021 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" event={"ID":"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8","Type":"ContainerStarted","Data":"4de08209cf3a8e98d800fca741e9c34256c9e6e0942a4b5f761e4336071af9ea"} Dec 01 22:31:17 crc kubenswrapper[4857]: I1201 22:31:17.898357 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" event={"ID":"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8","Type":"ContainerStarted","Data":"f812681677551a135d5996e9796404664ee0fada500992f9fa3c139a0cf341ea"} Dec 01 22:31:18 crc kubenswrapper[4857]: I1201 22:31:18.911964 4857 generic.go:334] "Generic (PLEG): container finished" podID="a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" containerID="4de08209cf3a8e98d800fca741e9c34256c9e6e0942a4b5f761e4336071af9ea" exitCode=0 Dec 01 22:31:18 crc kubenswrapper[4857]: I1201 22:31:18.912096 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" event={"ID":"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8","Type":"ContainerDied","Data":"4de08209cf3a8e98d800fca741e9c34256c9e6e0942a4b5f761e4336071af9ea"} Dec 01 22:31:20 crc kubenswrapper[4857]: I1201 22:31:20.942119 4857 generic.go:334] "Generic (PLEG): container finished" podID="a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" containerID="7efee4b6b81434fecc45daef666b7f2fb2986343d24f5cd663a91d8547956cdd" exitCode=0 Dec 01 22:31:20 crc kubenswrapper[4857]: I1201 22:31:20.942204 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" event={"ID":"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8","Type":"ContainerDied","Data":"7efee4b6b81434fecc45daef666b7f2fb2986343d24f5cd663a91d8547956cdd"} Dec 01 22:31:21 crc kubenswrapper[4857]: I1201 22:31:21.957399 4857 generic.go:334] "Generic (PLEG): container finished" podID="a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" containerID="4b3235e31d8db3e2840adb591f181e123815fde21c11e9060d894a398a1f3167" exitCode=0 Dec 01 22:31:21 crc kubenswrapper[4857]: I1201 22:31:21.957504 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" event={"ID":"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8","Type":"ContainerDied","Data":"4b3235e31d8db3e2840adb591f181e123815fde21c11e9060d894a398a1f3167"} Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.851837 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.864949 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dnsv\" (UniqueName: \"kubernetes.io/projected/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-kube-api-access-9dnsv\") pod \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\" (UID: \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\") " Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.865003 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-bundle\") pod \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\" (UID: \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\") " Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.865288 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-util\") pod \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\" (UID: \"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8\") " Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.868834 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-bundle" (OuterVolumeSpecName: "bundle") pod "a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" (UID: "a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.874994 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-kube-api-access-9dnsv" (OuterVolumeSpecName: "kube-api-access-9dnsv") pod "a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" (UID: "a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8"). InnerVolumeSpecName "kube-api-access-9dnsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.955067 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-util" (OuterVolumeSpecName: "util") pod "a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" (UID: "a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.968274 4857 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-util\") on node \"crc\" DevicePath \"\"" Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.968321 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dnsv\" (UniqueName: \"kubernetes.io/projected/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-kube-api-access-9dnsv\") on node \"crc\" DevicePath \"\"" Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.968340 4857 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.982898 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" event={"ID":"a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8","Type":"ContainerDied","Data":"f812681677551a135d5996e9796404664ee0fada500992f9fa3c139a0cf341ea"} Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.982942 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f812681677551a135d5996e9796404664ee0fada500992f9fa3c139a0cf341ea" Dec 01 22:31:23 crc kubenswrapper[4857]: I1201 22:31:23.982958 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd" Dec 01 22:31:28 crc kubenswrapper[4857]: I1201 22:31:28.835549 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:31:28 crc kubenswrapper[4857]: E1201 22:31:28.836553 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:31:40 crc kubenswrapper[4857]: I1201 22:31:40.835446 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:31:40 crc kubenswrapper[4857]: E1201 22:31:40.836415 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.538016 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586"] Dec 01 22:31:43 crc kubenswrapper[4857]: E1201 22:31:43.538969 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" containerName="extract" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.538983 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" containerName="extract" Dec 01 22:31:43 crc kubenswrapper[4857]: E1201 22:31:43.539012 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" containerName="util" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.539019 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" containerName="util" Dec 01 22:31:43 crc kubenswrapper[4857]: E1201 22:31:43.539031 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" containerName="pull" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.539050 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" containerName="pull" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.539246 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8" containerName="extract" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.539976 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.544623 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-l4665" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.549771 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.552105 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg"] Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.553409 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.565595 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586"] Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.575091 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-tqws8"] Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.576725 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-tqws8" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.581420 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.582123 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.584820 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg"] Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.587462 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-sf4dh" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.600882 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-tqws8"] Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.713723 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-bxmst"] Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.714925 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.716640 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.717026 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-2qchs" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.755831 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/203588e4-df3f-407e-9454-9cff35f281e5-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-744b77886-6c586\" (UID: \"203588e4-df3f-407e-9454-9cff35f281e5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.756149 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ncrd\" (UniqueName: \"kubernetes.io/projected/45d679d3-a9f8-402a-a8cf-4f9913ea40a1-kube-api-access-8ncrd\") pod \"obo-prometheus-operator-668cf9dfbb-tqws8\" (UID: \"45d679d3-a9f8-402a-a8cf-4f9913ea40a1\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-tqws8" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.756257 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e212d47c-24b3-44cb-a1de-3a291777358d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-744b77886-qczzg\" (UID: \"e212d47c-24b3-44cb-a1de-3a291777358d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.756394 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/203588e4-df3f-407e-9454-9cff35f281e5-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-744b77886-6c586\" (UID: \"203588e4-df3f-407e-9454-9cff35f281e5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.756474 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e212d47c-24b3-44cb-a1de-3a291777358d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-744b77886-qczzg\" (UID: \"e212d47c-24b3-44cb-a1de-3a291777358d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.764996 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-bxmst"] Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.858008 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/203588e4-df3f-407e-9454-9cff35f281e5-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-744b77886-6c586\" (UID: \"203588e4-df3f-407e-9454-9cff35f281e5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.858365 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ncrd\" (UniqueName: \"kubernetes.io/projected/45d679d3-a9f8-402a-a8cf-4f9913ea40a1-kube-api-access-8ncrd\") pod \"obo-prometheus-operator-668cf9dfbb-tqws8\" (UID: \"45d679d3-a9f8-402a-a8cf-4f9913ea40a1\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-tqws8" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.858606 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e212d47c-24b3-44cb-a1de-3a291777358d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-744b77886-qczzg\" (UID: \"e212d47c-24b3-44cb-a1de-3a291777358d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.858725 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-bxmst\" (UID: \"f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f\") " pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.859237 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/203588e4-df3f-407e-9454-9cff35f281e5-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-744b77886-6c586\" (UID: \"203588e4-df3f-407e-9454-9cff35f281e5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.859284 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e212d47c-24b3-44cb-a1de-3a291777358d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-744b77886-qczzg\" (UID: \"e212d47c-24b3-44cb-a1de-3a291777358d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.859314 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pl9w\" (UniqueName: \"kubernetes.io/projected/f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f-kube-api-access-7pl9w\") pod \"observability-operator-d8bb48f5d-bxmst\" (UID: \"f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f\") " pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.864118 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/203588e4-df3f-407e-9454-9cff35f281e5-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-744b77886-6c586\" (UID: \"203588e4-df3f-407e-9454-9cff35f281e5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.865464 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e212d47c-24b3-44cb-a1de-3a291777358d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-744b77886-qczzg\" (UID: \"e212d47c-24b3-44cb-a1de-3a291777358d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.868783 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e212d47c-24b3-44cb-a1de-3a291777358d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-744b77886-qczzg\" (UID: \"e212d47c-24b3-44cb-a1de-3a291777358d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.871981 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.891265 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ncrd\" (UniqueName: \"kubernetes.io/projected/45d679d3-a9f8-402a-a8cf-4f9913ea40a1-kube-api-access-8ncrd\") pod \"obo-prometheus-operator-668cf9dfbb-tqws8\" (UID: \"45d679d3-a9f8-402a-a8cf-4f9913ea40a1\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-tqws8" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.891557 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/203588e4-df3f-407e-9454-9cff35f281e5-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-744b77886-6c586\" (UID: \"203588e4-df3f-407e-9454-9cff35f281e5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.896555 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-tqws8" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.917785 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-26djf"] Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.923030 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-26djf" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.925698 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-tqfwm" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.935268 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-26djf"] Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.959986 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk62r\" (UniqueName: \"kubernetes.io/projected/f5a63bac-285c-401a-a975-3079cbad38cf-kube-api-access-wk62r\") pod \"perses-operator-5446b9c989-26djf\" (UID: \"f5a63bac-285c-401a-a975-3079cbad38cf\") " pod="openshift-operators/perses-operator-5446b9c989-26djf" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.960096 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-bxmst\" (UID: \"f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f\") " pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.960368 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pl9w\" (UniqueName: \"kubernetes.io/projected/f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f-kube-api-access-7pl9w\") pod \"observability-operator-d8bb48f5d-bxmst\" (UID: \"f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f\") " pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.960507 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f5a63bac-285c-401a-a975-3079cbad38cf-openshift-service-ca\") pod \"perses-operator-5446b9c989-26djf\" (UID: \"f5a63bac-285c-401a-a975-3079cbad38cf\") " pod="openshift-operators/perses-operator-5446b9c989-26djf" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.965608 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-bxmst\" (UID: \"f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f\") " pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" Dec 01 22:31:43 crc kubenswrapper[4857]: I1201 22:31:43.979067 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pl9w\" (UniqueName: \"kubernetes.io/projected/f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f-kube-api-access-7pl9w\") pod \"observability-operator-d8bb48f5d-bxmst\" (UID: \"f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f\") " pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" Dec 01 22:31:44 crc kubenswrapper[4857]: I1201 22:31:44.062831 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f5a63bac-285c-401a-a975-3079cbad38cf-openshift-service-ca\") pod \"perses-operator-5446b9c989-26djf\" (UID: \"f5a63bac-285c-401a-a975-3079cbad38cf\") " pod="openshift-operators/perses-operator-5446b9c989-26djf" Dec 01 22:31:44 crc kubenswrapper[4857]: I1201 22:31:44.062888 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk62r\" (UniqueName: \"kubernetes.io/projected/f5a63bac-285c-401a-a975-3079cbad38cf-kube-api-access-wk62r\") pod \"perses-operator-5446b9c989-26djf\" (UID: \"f5a63bac-285c-401a-a975-3079cbad38cf\") " pod="openshift-operators/perses-operator-5446b9c989-26djf" Dec 01 22:31:44 crc kubenswrapper[4857]: I1201 22:31:44.063973 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f5a63bac-285c-401a-a975-3079cbad38cf-openshift-service-ca\") pod \"perses-operator-5446b9c989-26djf\" (UID: \"f5a63bac-285c-401a-a975-3079cbad38cf\") " pod="openshift-operators/perses-operator-5446b9c989-26djf" Dec 01 22:31:44 crc kubenswrapper[4857]: I1201 22:31:44.076606 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" Dec 01 22:31:44 crc kubenswrapper[4857]: I1201 22:31:44.086647 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk62r\" (UniqueName: \"kubernetes.io/projected/f5a63bac-285c-401a-a975-3079cbad38cf-kube-api-access-wk62r\") pod \"perses-operator-5446b9c989-26djf\" (UID: \"f5a63bac-285c-401a-a975-3079cbad38cf\") " pod="openshift-operators/perses-operator-5446b9c989-26djf" Dec 01 22:31:44 crc kubenswrapper[4857]: I1201 22:31:44.162536 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586" Dec 01 22:31:44 crc kubenswrapper[4857]: I1201 22:31:44.372223 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-26djf" Dec 01 22:31:45 crc kubenswrapper[4857]: I1201 22:31:45.014593 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg"] Dec 01 22:31:45 crc kubenswrapper[4857]: I1201 22:31:45.084909 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586"] Dec 01 22:31:45 crc kubenswrapper[4857]: I1201 22:31:45.095033 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-bxmst"] Dec 01 22:31:45 crc kubenswrapper[4857]: I1201 22:31:45.102635 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-tqws8"] Dec 01 22:31:45 crc kubenswrapper[4857]: I1201 22:31:45.250496 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-26djf"] Dec 01 22:31:45 crc kubenswrapper[4857]: I1201 22:31:45.306096 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" event={"ID":"f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f","Type":"ContainerStarted","Data":"28262580b2c73f9e9fcd4e36c65018f367422548ac499a88623c17ae371b1d3c"} Dec 01 22:31:45 crc kubenswrapper[4857]: I1201 22:31:45.311243 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-tqws8" event={"ID":"45d679d3-a9f8-402a-a8cf-4f9913ea40a1","Type":"ContainerStarted","Data":"bb5670fed9333030f7cef42a38178d7e965567e98e446a59a73959e0a645b22a"} Dec 01 22:31:45 crc kubenswrapper[4857]: I1201 22:31:45.316745 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-26djf" event={"ID":"f5a63bac-285c-401a-a975-3079cbad38cf","Type":"ContainerStarted","Data":"ff145202f5e7c626be0e3f1f7606ec8b622ce970b63240287a340acd4f42cbdd"} Dec 01 22:31:45 crc kubenswrapper[4857]: I1201 22:31:45.320807 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586" event={"ID":"203588e4-df3f-407e-9454-9cff35f281e5","Type":"ContainerStarted","Data":"db162de1b75d29b5614891bdb16391302d2115bec13398712571fe2cf3ccc5de"} Dec 01 22:31:45 crc kubenswrapper[4857]: I1201 22:31:45.322756 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg" event={"ID":"e212d47c-24b3-44cb-a1de-3a291777358d","Type":"ContainerStarted","Data":"f8c4c284f4fc306be53f2d6ab2f00fcffaad32ebb63efc928933dbdb90344ba4"} Dec 01 22:31:54 crc kubenswrapper[4857]: I1201 22:31:54.835875 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:31:54 crc kubenswrapper[4857]: E1201 22:31:54.836594 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:32:00 crc kubenswrapper[4857]: E1201 22:32:00.946798 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385" Dec 01 22:32:00 crc kubenswrapper[4857]: E1201 22:32:00.947282 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wk62r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-5446b9c989-26djf_openshift-operators(f5a63bac-285c-401a-a975-3079cbad38cf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 22:32:00 crc kubenswrapper[4857]: E1201 22:32:00.948526 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-5446b9c989-26djf" podUID="f5a63bac-285c-401a-a975-3079cbad38cf" Dec 01 22:32:01 crc kubenswrapper[4857]: E1201 22:32:01.009149 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 01 22:32:01 crc kubenswrapper[4857]: E1201 22:32:01.009323 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-744b77886-qczzg_openshift-operators(e212d47c-24b3-44cb-a1de-3a291777358d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 22:32:01 crc kubenswrapper[4857]: E1201 22:32:01.011374 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg" podUID="e212d47c-24b3-44cb-a1de-3a291777358d" Dec 01 22:32:01 crc kubenswrapper[4857]: I1201 22:32:01.214027 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 22:32:01 crc kubenswrapper[4857]: I1201 22:32:01.222144 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-api" containerID="cri-o://3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3" gracePeriod=30 Dec 01 22:32:01 crc kubenswrapper[4857]: I1201 22:32:01.223458 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-listener" containerID="cri-o://ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee" gracePeriod=30 Dec 01 22:32:01 crc kubenswrapper[4857]: I1201 22:32:01.223526 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-notifier" containerID="cri-o://bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc" gracePeriod=30 Dec 01 22:32:01 crc kubenswrapper[4857]: I1201 22:32:01.223563 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-evaluator" containerID="cri-o://641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e" gracePeriod=30 Dec 01 22:32:01 crc kubenswrapper[4857]: E1201 22:32:01.934772 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385\\\"\"" pod="openshift-operators/perses-operator-5446b9c989-26djf" podUID="f5a63bac-285c-401a-a975-3079cbad38cf" Dec 01 22:32:02 crc kubenswrapper[4857]: I1201 22:32:02.943416 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" event={"ID":"f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f","Type":"ContainerStarted","Data":"807e3603d2db9832ab518daabf8ef648ccb55cee88532d1896ef37d9bf6a23a9"} Dec 01 22:32:02 crc kubenswrapper[4857]: I1201 22:32:02.943923 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" Dec 01 22:32:02 crc kubenswrapper[4857]: I1201 22:32:02.945187 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-tqws8" event={"ID":"45d679d3-a9f8-402a-a8cf-4f9913ea40a1","Type":"ContainerStarted","Data":"b11a3b6ae60f35d9f0c0f759c1ff2a7c6af58746b2fe906b2ed21a5e986649b9"} Dec 01 22:32:02 crc kubenswrapper[4857]: I1201 22:32:02.945302 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" Dec 01 22:32:02 crc kubenswrapper[4857]: I1201 22:32:02.946877 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586" event={"ID":"203588e4-df3f-407e-9454-9cff35f281e5","Type":"ContainerStarted","Data":"e15c6311bd64941b9b49a1192d5b740c5a043a7e6b1b40f227ff7e58301977ec"} Dec 01 22:32:02 crc kubenswrapper[4857]: I1201 22:32:02.955147 4857 generic.go:334] "Generic (PLEG): container finished" podID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerID="641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e" exitCode=0 Dec 01 22:32:02 crc kubenswrapper[4857]: I1201 22:32:02.955179 4857 generic.go:334] "Generic (PLEG): container finished" podID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerID="3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3" exitCode=0 Dec 01 22:32:02 crc kubenswrapper[4857]: I1201 22:32:02.955231 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e8c00e5e-1174-4498-8242-3b6c5b765f00","Type":"ContainerDied","Data":"641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e"} Dec 01 22:32:02 crc kubenswrapper[4857]: I1201 22:32:02.955255 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e8c00e5e-1174-4498-8242-3b6c5b765f00","Type":"ContainerDied","Data":"3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3"} Dec 01 22:32:02 crc kubenswrapper[4857]: I1201 22:32:02.958611 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg" event={"ID":"e212d47c-24b3-44cb-a1de-3a291777358d","Type":"ContainerStarted","Data":"2b3e3400ece0646760cdcd8c26466a6713afc736e2b08b2cf86e90009b461f1a"} Dec 01 22:32:02 crc kubenswrapper[4857]: I1201 22:32:02.972132 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-bxmst" podStartSLOduration=4.023855687 podStartE2EDuration="19.972107694s" podCreationTimestamp="2025-12-01 22:31:43 +0000 UTC" firstStartedPulling="2025-12-01 22:31:45.108523186 +0000 UTC m=+3423.598585503" lastFinishedPulling="2025-12-01 22:32:01.056775193 +0000 UTC m=+3439.546837510" observedRunningTime="2025-12-01 22:32:02.97074093 +0000 UTC m=+3441.460803257" watchObservedRunningTime="2025-12-01 22:32:02.972107694 +0000 UTC m=+3441.462170011" Dec 01 22:32:03 crc kubenswrapper[4857]: I1201 22:32:03.004684 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-6c586" podStartSLOduration=4.056838756 podStartE2EDuration="20.004664373s" podCreationTimestamp="2025-12-01 22:31:43 +0000 UTC" firstStartedPulling="2025-12-01 22:31:45.108945806 +0000 UTC m=+3423.599008123" lastFinishedPulling="2025-12-01 22:32:01.056771423 +0000 UTC m=+3439.546833740" observedRunningTime="2025-12-01 22:32:02.999992168 +0000 UTC m=+3441.490054485" watchObservedRunningTime="2025-12-01 22:32:03.004664373 +0000 UTC m=+3441.494726690" Dec 01 22:32:03 crc kubenswrapper[4857]: I1201 22:32:03.398291 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-744b77886-qczzg" podStartSLOduration=-9223372016.456509 podStartE2EDuration="20.398267953s" podCreationTimestamp="2025-12-01 22:31:43 +0000 UTC" firstStartedPulling="2025-12-01 22:31:45.025014416 +0000 UTC m=+3423.515076723" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 22:32:03.338973397 +0000 UTC m=+3441.829035714" watchObservedRunningTime="2025-12-01 22:32:03.398267953 +0000 UTC m=+3441.888330270" Dec 01 22:32:03 crc kubenswrapper[4857]: I1201 22:32:03.424719 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-tqws8" podStartSLOduration=4.411435588 podStartE2EDuration="20.424695101s" podCreationTimestamp="2025-12-01 22:31:43 +0000 UTC" firstStartedPulling="2025-12-01 22:31:45.109693434 +0000 UTC m=+3423.599755751" lastFinishedPulling="2025-12-01 22:32:01.122952947 +0000 UTC m=+3439.613015264" observedRunningTime="2025-12-01 22:32:03.376780505 +0000 UTC m=+3441.866842822" watchObservedRunningTime="2025-12-01 22:32:03.424695101 +0000 UTC m=+3441.914757418" Dec 01 22:32:03 crc kubenswrapper[4857]: I1201 22:32:03.967820 4857 generic.go:334] "Generic (PLEG): container finished" podID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerID="ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee" exitCode=0 Dec 01 22:32:03 crc kubenswrapper[4857]: I1201 22:32:03.968597 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e8c00e5e-1174-4498-8242-3b6c5b765f00","Type":"ContainerDied","Data":"ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee"} Dec 01 22:32:05 crc kubenswrapper[4857]: I1201 22:32:05.986335 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 01 22:32:05 crc kubenswrapper[4857]: I1201 22:32:05.988575 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:05 crc kubenswrapper[4857]: I1201 22:32:05.991846 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 01 22:32:05 crc kubenswrapper[4857]: I1201 22:32:05.991994 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 01 22:32:05 crc kubenswrapper[4857]: I1201 22:32:05.992078 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-8jknm" Dec 01 22:32:05 crc kubenswrapper[4857]: I1201 22:32:05.992180 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 01 22:32:05 crc kubenswrapper[4857]: I1201 22:32:05.992217 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.002082 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.081376 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.081511 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.081558 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.081585 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.081639 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpvgk\" (UniqueName: \"kubernetes.io/projected/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-kube-api-access-gpvgk\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.081666 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.081738 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.183861 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.183946 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.183981 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.184097 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.185001 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.185031 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.185098 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpvgk\" (UniqueName: \"kubernetes.io/projected/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-kube-api-access-gpvgk\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.185644 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.191357 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.191910 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.192825 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.193208 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.212172 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpvgk\" (UniqueName: \"kubernetes.io/projected/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-kube-api-access-gpvgk\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.219111 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/98dbb0ff-4644-431e-801e-5ce7fb3f6f81-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"98dbb0ff-4644-431e-801e-5ce7fb3f6f81\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:06 crc kubenswrapper[4857]: I1201 22:32:06.364217 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:06.837243 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.241265 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.243761 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.246101 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.246367 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-whwdt" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.246526 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.246714 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.249331 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.266449 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.281772 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.371150 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.371294 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.371342 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/28a6d05a-72fa-4f72-a3d1-d6891412c04f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.371376 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.371402 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-config\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.371576 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/28a6d05a-72fa-4f72-a3d1-d6891412c04f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.371636 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/28a6d05a-72fa-4f72-a3d1-d6891412c04f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.371676 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92ngc\" (UniqueName: \"kubernetes.io/projected/28a6d05a-72fa-4f72-a3d1-d6891412c04f-kube-api-access-92ngc\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.474236 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.474284 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/28a6d05a-72fa-4f72-a3d1-d6891412c04f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.474309 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.474331 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-config\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.474393 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/28a6d05a-72fa-4f72-a3d1-d6891412c04f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.474415 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/28a6d05a-72fa-4f72-a3d1-d6891412c04f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.474432 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92ngc\" (UniqueName: \"kubernetes.io/projected/28a6d05a-72fa-4f72-a3d1-d6891412c04f-kube-api-access-92ngc\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.474531 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.474937 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.477600 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/28a6d05a-72fa-4f72-a3d1-d6891412c04f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.481680 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.481748 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-config\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.483233 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/28a6d05a-72fa-4f72-a3d1-d6891412c04f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.499147 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/28a6d05a-72fa-4f72-a3d1-d6891412c04f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.499267 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.503758 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92ngc\" (UniqueName: \"kubernetes.io/projected/28a6d05a-72fa-4f72-a3d1-d6891412c04f-kube-api-access-92ngc\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.510802 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:07 crc kubenswrapper[4857]: I1201 22:32:07.603485 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:08 crc kubenswrapper[4857]: I1201 22:32:08.088315 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"7018e5ef4fa6765cca44e184b68c5e2a16e99c3def48589b77a1494d6ddc3ee9"} Dec 01 22:32:08 crc kubenswrapper[4857]: I1201 22:32:08.305254 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 01 22:32:08 crc kubenswrapper[4857]: I1201 22:32:08.368232 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:32:09 crc kubenswrapper[4857]: I1201 22:32:09.101849 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"28a6d05a-72fa-4f72-a3d1-d6891412c04f","Type":"ContainerStarted","Data":"558a5f392b336c53e5059f355db55af8e3ef04312ef00e0ba2474404bc9c3f21"} Dec 01 22:32:09 crc kubenswrapper[4857]: I1201 22:32:09.104866 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"98dbb0ff-4644-431e-801e-5ce7fb3f6f81","Type":"ContainerStarted","Data":"071efc8119971face9ff8d0fd06d468d2f249088105d1b69d18a9fc37c28dbd7"} Dec 01 22:32:11 crc kubenswrapper[4857]: I1201 22:32:11.938384 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.040619 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-scripts\") pod \"e8c00e5e-1174-4498-8242-3b6c5b765f00\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.040724 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-config-data\") pod \"e8c00e5e-1174-4498-8242-3b6c5b765f00\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.040776 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-public-tls-certs\") pod \"e8c00e5e-1174-4498-8242-3b6c5b765f00\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.040839 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-combined-ca-bundle\") pod \"e8c00e5e-1174-4498-8242-3b6c5b765f00\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.040905 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtgvt\" (UniqueName: \"kubernetes.io/projected/e8c00e5e-1174-4498-8242-3b6c5b765f00-kube-api-access-xtgvt\") pod \"e8c00e5e-1174-4498-8242-3b6c5b765f00\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.040933 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-internal-tls-certs\") pod \"e8c00e5e-1174-4498-8242-3b6c5b765f00\" (UID: \"e8c00e5e-1174-4498-8242-3b6c5b765f00\") " Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.055278 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-scripts" (OuterVolumeSpecName: "scripts") pod "e8c00e5e-1174-4498-8242-3b6c5b765f00" (UID: "e8c00e5e-1174-4498-8242-3b6c5b765f00"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.067355 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8c00e5e-1174-4498-8242-3b6c5b765f00-kube-api-access-xtgvt" (OuterVolumeSpecName: "kube-api-access-xtgvt") pod "e8c00e5e-1174-4498-8242-3b6c5b765f00" (UID: "e8c00e5e-1174-4498-8242-3b6c5b765f00"). InnerVolumeSpecName "kube-api-access-xtgvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.151231 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtgvt\" (UniqueName: \"kubernetes.io/projected/e8c00e5e-1174-4498-8242-3b6c5b765f00-kube-api-access-xtgvt\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.151256 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.183430 4857 generic.go:334] "Generic (PLEG): container finished" podID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerID="bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc" exitCode=0 Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.183474 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e8c00e5e-1174-4498-8242-3b6c5b765f00","Type":"ContainerDied","Data":"bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc"} Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.183505 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e8c00e5e-1174-4498-8242-3b6c5b765f00","Type":"ContainerDied","Data":"404f22886916e56c907db5f2753cbed85d04ed13ff5ee5ac2157fb56c676f8fd"} Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.183525 4857 scope.go:117] "RemoveContainer" containerID="ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.183693 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.198234 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e8c00e5e-1174-4498-8242-3b6c5b765f00" (UID: "e8c00e5e-1174-4498-8242-3b6c5b765f00"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.224131 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e8c00e5e-1174-4498-8242-3b6c5b765f00" (UID: "e8c00e5e-1174-4498-8242-3b6c5b765f00"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.254977 4857 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.255010 4857 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.259552 4857 scope.go:117] "RemoveContainer" containerID="bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.266269 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-config-data" (OuterVolumeSpecName: "config-data") pod "e8c00e5e-1174-4498-8242-3b6c5b765f00" (UID: "e8c00e5e-1174-4498-8242-3b6c5b765f00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.278236 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8c00e5e-1174-4498-8242-3b6c5b765f00" (UID: "e8c00e5e-1174-4498-8242-3b6c5b765f00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.282690 4857 scope.go:117] "RemoveContainer" containerID="641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.301509 4857 scope.go:117] "RemoveContainer" containerID="3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.320705 4857 scope.go:117] "RemoveContainer" containerID="ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee" Dec 01 22:32:12 crc kubenswrapper[4857]: E1201 22:32:12.321174 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee\": container with ID starting with ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee not found: ID does not exist" containerID="ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.321217 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee"} err="failed to get container status \"ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee\": rpc error: code = NotFound desc = could not find container \"ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee\": container with ID starting with ff21005506441674397d37ec3cbaa04bf839bc03b3c704bf92830c8550c885ee not found: ID does not exist" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.321243 4857 scope.go:117] "RemoveContainer" containerID="bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc" Dec 01 22:32:12 crc kubenswrapper[4857]: E1201 22:32:12.321593 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc\": container with ID starting with bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc not found: ID does not exist" containerID="bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.321637 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc"} err="failed to get container status \"bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc\": rpc error: code = NotFound desc = could not find container \"bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc\": container with ID starting with bc1f60bc0144f21858ebaa135d729b4cc81baa3ea5e5fe031468fa0f80ec70fc not found: ID does not exist" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.321667 4857 scope.go:117] "RemoveContainer" containerID="641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e" Dec 01 22:32:12 crc kubenswrapper[4857]: E1201 22:32:12.321906 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e\": container with ID starting with 641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e not found: ID does not exist" containerID="641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.321928 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e"} err="failed to get container status \"641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e\": rpc error: code = NotFound desc = could not find container \"641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e\": container with ID starting with 641c033d24e7301697d7678bd0afa47c1c49e13b5960a7e5aa01047e65c8c93e not found: ID does not exist" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.321940 4857 scope.go:117] "RemoveContainer" containerID="3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3" Dec 01 22:32:12 crc kubenswrapper[4857]: E1201 22:32:12.322178 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3\": container with ID starting with 3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3 not found: ID does not exist" containerID="3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.322208 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3"} err="failed to get container status \"3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3\": rpc error: code = NotFound desc = could not find container \"3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3\": container with ID starting with 3c8a5accb4e52dbbc424d52c33be8dd2f34bcb594aec3dd8d0457bef637ae4f3 not found: ID does not exist" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.356818 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.356851 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c00e5e-1174-4498-8242-3b6c5b765f00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.520432 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.538098 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.550081 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 01 22:32:12 crc kubenswrapper[4857]: E1201 22:32:12.550749 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-listener" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.550783 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-listener" Dec 01 22:32:12 crc kubenswrapper[4857]: E1201 22:32:12.550844 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-notifier" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.550864 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-notifier" Dec 01 22:32:12 crc kubenswrapper[4857]: E1201 22:32:12.550905 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-evaluator" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.550925 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-evaluator" Dec 01 22:32:12 crc kubenswrapper[4857]: E1201 22:32:12.550948 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-api" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.550960 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-api" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.551337 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-listener" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.551371 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-evaluator" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.551390 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-notifier" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.551420 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" containerName="aodh-api" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.554895 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.559948 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.601456 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-j6fzf" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.601503 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.601775 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.601924 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.602084 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.667459 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-internal-tls-certs\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.667502 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-config-data\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.667567 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-public-tls-certs\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.667601 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8xdw\" (UniqueName: \"kubernetes.io/projected/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-kube-api-access-r8xdw\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.667679 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-scripts\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.667720 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.769372 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.769462 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-internal-tls-certs\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.769494 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-config-data\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.769559 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-public-tls-certs\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.769597 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8xdw\" (UniqueName: \"kubernetes.io/projected/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-kube-api-access-r8xdw\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.769637 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-scripts\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.773751 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.775389 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-config-data\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.777666 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-scripts\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.781169 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-internal-tls-certs\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.796926 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8xdw\" (UniqueName: \"kubernetes.io/projected/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-kube-api-access-r8xdw\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.857754 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-public-tls-certs\") pod \"aodh-0\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " pod="openstack/aodh-0" Dec 01 22:32:12 crc kubenswrapper[4857]: I1201 22:32:12.945224 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:32:13 crc kubenswrapper[4857]: I1201 22:32:13.565099 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 22:32:13 crc kubenswrapper[4857]: W1201 22:32:13.764867 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2d444dd_0c4b_48ed_927f_9e12b80b9c26.slice/crio-4c8b4d2fb1d2f5ed4bc544eb5eca7c6f8ce41c5b8d35039cf8e75db4672d810b WatchSource:0}: Error finding container 4c8b4d2fb1d2f5ed4bc544eb5eca7c6f8ce41c5b8d35039cf8e75db4672d810b: Status 404 returned error can't find the container with id 4c8b4d2fb1d2f5ed4bc544eb5eca7c6f8ce41c5b8d35039cf8e75db4672d810b Dec 01 22:32:13 crc kubenswrapper[4857]: I1201 22:32:13.868538 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8c00e5e-1174-4498-8242-3b6c5b765f00" path="/var/lib/kubelet/pods/e8c00e5e-1174-4498-8242-3b6c5b765f00/volumes" Dec 01 22:32:14 crc kubenswrapper[4857]: I1201 22:32:14.342121 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d2d444dd-0c4b-48ed-927f-9e12b80b9c26","Type":"ContainerStarted","Data":"4c8b4d2fb1d2f5ed4bc544eb5eca7c6f8ce41c5b8d35039cf8e75db4672d810b"} Dec 01 22:32:15 crc kubenswrapper[4857]: I1201 22:32:15.395395 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"98dbb0ff-4644-431e-801e-5ce7fb3f6f81","Type":"ContainerStarted","Data":"1fc7fc7810b48261d566e22627385e271382578646c0f27fd9c0d88ab9092370"} Dec 01 22:32:15 crc kubenswrapper[4857]: I1201 22:32:15.397407 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d2d444dd-0c4b-48ed-927f-9e12b80b9c26","Type":"ContainerStarted","Data":"e7928e5378f29db5337410c2925241367e1f79a6f6f038ad6b2966156971d56f"} Dec 01 22:32:15 crc kubenswrapper[4857]: I1201 22:32:15.400824 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"28a6d05a-72fa-4f72-a3d1-d6891412c04f","Type":"ContainerStarted","Data":"7eb7848446e6cecd845451a88545310fcff1786bb62581ca40a7411832c137c1"} Dec 01 22:32:16 crc kubenswrapper[4857]: I1201 22:32:16.462640 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d2d444dd-0c4b-48ed-927f-9e12b80b9c26","Type":"ContainerStarted","Data":"3ecb47734c66430237ae9b314bd49ab2c834939b9edd79577cb35f70fab19150"} Dec 01 22:32:17 crc kubenswrapper[4857]: I1201 22:32:17.480477 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-26djf" event={"ID":"f5a63bac-285c-401a-a975-3079cbad38cf","Type":"ContainerStarted","Data":"a97be0f327966cc6d001aa6bd28dcdfe3a10bdddddef0190668030af0d163092"} Dec 01 22:32:17 crc kubenswrapper[4857]: I1201 22:32:17.482369 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-26djf" Dec 01 22:32:17 crc kubenswrapper[4857]: I1201 22:32:17.487235 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d2d444dd-0c4b-48ed-927f-9e12b80b9c26","Type":"ContainerStarted","Data":"7f0cbeb943e426c39f61ced53e6f1266fc9a5661fa554d4a48415740ecd5d669"} Dec 01 22:32:17 crc kubenswrapper[4857]: I1201 22:32:17.512306 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-26djf" podStartSLOduration=3.581224342 podStartE2EDuration="34.512284881s" podCreationTimestamp="2025-12-01 22:31:43 +0000 UTC" firstStartedPulling="2025-12-01 22:31:45.238596658 +0000 UTC m=+3423.728658975" lastFinishedPulling="2025-12-01 22:32:16.169657177 +0000 UTC m=+3454.659719514" observedRunningTime="2025-12-01 22:32:17.503813533 +0000 UTC m=+3455.993875850" watchObservedRunningTime="2025-12-01 22:32:17.512284881 +0000 UTC m=+3456.002347198" Dec 01 22:32:18 crc kubenswrapper[4857]: I1201 22:32:18.502414 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d2d444dd-0c4b-48ed-927f-9e12b80b9c26","Type":"ContainerStarted","Data":"5144a43c6f9c4eada7f058dcd0cbb90d3df519122f7d62fafebdd27443472174"} Dec 01 22:32:18 crc kubenswrapper[4857]: I1201 22:32:18.526076 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.04080399 podStartE2EDuration="6.526058383s" podCreationTimestamp="2025-12-01 22:32:12 +0000 UTC" firstStartedPulling="2025-12-01 22:32:13.769446406 +0000 UTC m=+3452.259508753" lastFinishedPulling="2025-12-01 22:32:17.254700809 +0000 UTC m=+3455.744763146" observedRunningTime="2025-12-01 22:32:18.525016017 +0000 UTC m=+3457.015078384" watchObservedRunningTime="2025-12-01 22:32:18.526058383 +0000 UTC m=+3457.016120700" Dec 01 22:32:22 crc kubenswrapper[4857]: I1201 22:32:22.557927 4857 generic.go:334] "Generic (PLEG): container finished" podID="98dbb0ff-4644-431e-801e-5ce7fb3f6f81" containerID="1fc7fc7810b48261d566e22627385e271382578646c0f27fd9c0d88ab9092370" exitCode=0 Dec 01 22:32:22 crc kubenswrapper[4857]: I1201 22:32:22.558026 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"98dbb0ff-4644-431e-801e-5ce7fb3f6f81","Type":"ContainerDied","Data":"1fc7fc7810b48261d566e22627385e271382578646c0f27fd9c0d88ab9092370"} Dec 01 22:32:23 crc kubenswrapper[4857]: I1201 22:32:23.573318 4857 generic.go:334] "Generic (PLEG): container finished" podID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerID="7eb7848446e6cecd845451a88545310fcff1786bb62581ca40a7411832c137c1" exitCode=0 Dec 01 22:32:23 crc kubenswrapper[4857]: I1201 22:32:23.573398 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"28a6d05a-72fa-4f72-a3d1-d6891412c04f","Type":"ContainerDied","Data":"7eb7848446e6cecd845451a88545310fcff1786bb62581ca40a7411832c137c1"} Dec 01 22:32:24 crc kubenswrapper[4857]: I1201 22:32:24.375875 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-26djf" Dec 01 22:32:25 crc kubenswrapper[4857]: I1201 22:32:25.605022 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"98dbb0ff-4644-431e-801e-5ce7fb3f6f81","Type":"ContainerStarted","Data":"04df675d9822c689e3ab7ac9e97b07f41523387f6a34d987dfdab4948d1f02a9"} Dec 01 22:32:30 crc kubenswrapper[4857]: I1201 22:32:30.688487 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"98dbb0ff-4644-431e-801e-5ce7fb3f6f81","Type":"ContainerStarted","Data":"675cbc36aff86a3ddaa5d63466427d0b0c353f5519a7f076d68938ae99242e2d"} Dec 01 22:32:30 crc kubenswrapper[4857]: I1201 22:32:30.690412 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:30 crc kubenswrapper[4857]: I1201 22:32:30.693716 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 01 22:32:30 crc kubenswrapper[4857]: I1201 22:32:30.721338 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=8.770893023 podStartE2EDuration="25.721309137s" podCreationTimestamp="2025-12-01 22:32:05 +0000 UTC" firstStartedPulling="2025-12-01 22:32:08.321928081 +0000 UTC m=+3446.811990398" lastFinishedPulling="2025-12-01 22:32:25.272344195 +0000 UTC m=+3463.762406512" observedRunningTime="2025-12-01 22:32:30.719495282 +0000 UTC m=+3469.209557639" watchObservedRunningTime="2025-12-01 22:32:30.721309137 +0000 UTC m=+3469.211371494" Dec 01 22:32:33 crc kubenswrapper[4857]: I1201 22:32:33.723556 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"28a6d05a-72fa-4f72-a3d1-d6891412c04f","Type":"ContainerStarted","Data":"d4e54e17db288345bb6a6d7443248544c5e620caa42354563068faf1ae1f4ebd"} Dec 01 22:32:38 crc kubenswrapper[4857]: I1201 22:32:38.787770 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"28a6d05a-72fa-4f72-a3d1-d6891412c04f","Type":"ContainerStarted","Data":"0c4f7feff9667adcf34feb5ec56a6ee4ca861aaa9269eb8bb7c66bf6e117df5b"} Dec 01 22:32:42 crc kubenswrapper[4857]: I1201 22:32:42.831552 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"28a6d05a-72fa-4f72-a3d1-d6891412c04f","Type":"ContainerStarted","Data":"dc12ccdb703732ffdb5d672243e011945bc1097877909e725ab650bbdd457589"} Dec 01 22:32:42 crc kubenswrapper[4857]: I1201 22:32:42.873814 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.705280424 podStartE2EDuration="36.873781099s" podCreationTimestamp="2025-12-01 22:32:06 +0000 UTC" firstStartedPulling="2025-12-01 22:32:08.474918246 +0000 UTC m=+3446.964980563" lastFinishedPulling="2025-12-01 22:32:41.643418891 +0000 UTC m=+3480.133481238" observedRunningTime="2025-12-01 22:32:42.866112201 +0000 UTC m=+3481.356174548" watchObservedRunningTime="2025-12-01 22:32:42.873781099 +0000 UTC m=+3481.363843466" Dec 01 22:32:47 crc kubenswrapper[4857]: I1201 22:32:47.688361 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:52 crc kubenswrapper[4857]: I1201 22:32:52.602824 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:52 crc kubenswrapper[4857]: I1201 22:32:52.609691 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:52 crc kubenswrapper[4857]: I1201 22:32:52.948941 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.639987 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.640875 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" containerName="openstackclient" containerID="cri-o://134f2bba3e91ccb269c800ef4cdd79e63cf2a734297141290056902d1d9e4c57" gracePeriod=2 Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.655198 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.689813 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 22:32:54 crc kubenswrapper[4857]: E1201 22:32:54.690556 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" containerName="openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.690589 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" containerName="openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.691025 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" containerName="openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.692365 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.705242 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.720186 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" podUID="8b715e70-6964-4dd3-9c88-67d41fd7ad7a" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.757163 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nb7g\" (UniqueName: \"kubernetes.io/projected/8b715e70-6964-4dd3-9c88-67d41fd7ad7a-kube-api-access-5nb7g\") pod \"openstackclient\" (UID: \"8b715e70-6964-4dd3-9c88-67d41fd7ad7a\") " pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.757410 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b715e70-6964-4dd3-9c88-67d41fd7ad7a-openstack-config-secret\") pod \"openstackclient\" (UID: \"8b715e70-6964-4dd3-9c88-67d41fd7ad7a\") " pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.757572 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b715e70-6964-4dd3-9c88-67d41fd7ad7a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8b715e70-6964-4dd3-9c88-67d41fd7ad7a\") " pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.757629 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b715e70-6964-4dd3-9c88-67d41fd7ad7a-openstack-config\") pod \"openstackclient\" (UID: \"8b715e70-6964-4dd3-9c88-67d41fd7ad7a\") " pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.859593 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b715e70-6964-4dd3-9c88-67d41fd7ad7a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8b715e70-6964-4dd3-9c88-67d41fd7ad7a\") " pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.859943 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b715e70-6964-4dd3-9c88-67d41fd7ad7a-openstack-config\") pod \"openstackclient\" (UID: \"8b715e70-6964-4dd3-9c88-67d41fd7ad7a\") " pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.860032 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nb7g\" (UniqueName: \"kubernetes.io/projected/8b715e70-6964-4dd3-9c88-67d41fd7ad7a-kube-api-access-5nb7g\") pod \"openstackclient\" (UID: \"8b715e70-6964-4dd3-9c88-67d41fd7ad7a\") " pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.860192 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b715e70-6964-4dd3-9c88-67d41fd7ad7a-openstack-config-secret\") pod \"openstackclient\" (UID: \"8b715e70-6964-4dd3-9c88-67d41fd7ad7a\") " pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.861017 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b715e70-6964-4dd3-9c88-67d41fd7ad7a-openstack-config\") pod \"openstackclient\" (UID: \"8b715e70-6964-4dd3-9c88-67d41fd7ad7a\") " pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.865547 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b715e70-6964-4dd3-9c88-67d41fd7ad7a-openstack-config-secret\") pod \"openstackclient\" (UID: \"8b715e70-6964-4dd3-9c88-67d41fd7ad7a\") " pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.865567 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b715e70-6964-4dd3-9c88-67d41fd7ad7a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8b715e70-6964-4dd3-9c88-67d41fd7ad7a\") " pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.881743 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nb7g\" (UniqueName: \"kubernetes.io/projected/8b715e70-6964-4dd3-9c88-67d41fd7ad7a-kube-api-access-5nb7g\") pod \"openstackclient\" (UID: \"8b715e70-6964-4dd3-9c88-67d41fd7ad7a\") " pod="openstack/openstackclient" Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.964596 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.965402 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-api" containerID="cri-o://e7928e5378f29db5337410c2925241367e1f79a6f6f038ad6b2966156971d56f" gracePeriod=30 Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.965514 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-listener" containerID="cri-o://5144a43c6f9c4eada7f058dcd0cbb90d3df519122f7d62fafebdd27443472174" gracePeriod=30 Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.965538 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-evaluator" containerID="cri-o://3ecb47734c66430237ae9b314bd49ab2c834939b9edd79577cb35f70fab19150" gracePeriod=30 Dec 01 22:32:54 crc kubenswrapper[4857]: I1201 22:32:54.965538 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-notifier" containerID="cri-o://7f0cbeb943e426c39f61ced53e6f1266fc9a5661fa554d4a48415740ecd5d669" gracePeriod=30 Dec 01 22:32:55 crc kubenswrapper[4857]: I1201 22:32:55.026605 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 22:32:55 crc kubenswrapper[4857]: W1201 22:32:55.569831 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b715e70_6964_4dd3_9c88_67d41fd7ad7a.slice/crio-96f494f8264f9c92e08ceefae91596605356ad6a58edf091a9f5d5c735c4a2e6 WatchSource:0}: Error finding container 96f494f8264f9c92e08ceefae91596605356ad6a58edf091a9f5d5c735c4a2e6: Status 404 returned error can't find the container with id 96f494f8264f9c92e08ceefae91596605356ad6a58edf091a9f5d5c735c4a2e6 Dec 01 22:32:55 crc kubenswrapper[4857]: I1201 22:32:55.572921 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 22:32:55 crc kubenswrapper[4857]: I1201 22:32:55.975140 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"8b715e70-6964-4dd3-9c88-67d41fd7ad7a","Type":"ContainerStarted","Data":"c0400128b40261d57a1175c8f38e6fc05b2fe85dcd734a347555fcb3dfe80f93"} Dec 01 22:32:55 crc kubenswrapper[4857]: I1201 22:32:55.975512 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"8b715e70-6964-4dd3-9c88-67d41fd7ad7a","Type":"ContainerStarted","Data":"96f494f8264f9c92e08ceefae91596605356ad6a58edf091a9f5d5c735c4a2e6"} Dec 01 22:32:55 crc kubenswrapper[4857]: I1201 22:32:55.977593 4857 generic.go:334] "Generic (PLEG): container finished" podID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerID="3ecb47734c66430237ae9b314bd49ab2c834939b9edd79577cb35f70fab19150" exitCode=0 Dec 01 22:32:55 crc kubenswrapper[4857]: I1201 22:32:55.977619 4857 generic.go:334] "Generic (PLEG): container finished" podID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerID="e7928e5378f29db5337410c2925241367e1f79a6f6f038ad6b2966156971d56f" exitCode=0 Dec 01 22:32:55 crc kubenswrapper[4857]: I1201 22:32:55.977639 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d2d444dd-0c4b-48ed-927f-9e12b80b9c26","Type":"ContainerDied","Data":"3ecb47734c66430237ae9b314bd49ab2c834939b9edd79577cb35f70fab19150"} Dec 01 22:32:55 crc kubenswrapper[4857]: I1201 22:32:55.977662 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d2d444dd-0c4b-48ed-927f-9e12b80b9c26","Type":"ContainerDied","Data":"e7928e5378f29db5337410c2925241367e1f79a6f6f038ad6b2966156971d56f"} Dec 01 22:32:55 crc kubenswrapper[4857]: I1201 22:32:55.992578 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.9925588699999999 podStartE2EDuration="1.99255887s" podCreationTimestamp="2025-12-01 22:32:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 22:32:55.989977647 +0000 UTC m=+3494.480039954" watchObservedRunningTime="2025-12-01 22:32:55.99255887 +0000 UTC m=+3494.482621187" Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.304347 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.304609 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="prometheus" containerID="cri-o://d4e54e17db288345bb6a6d7443248544c5e620caa42354563068faf1ae1f4ebd" gracePeriod=600 Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.304754 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="thanos-sidecar" containerID="cri-o://dc12ccdb703732ffdb5d672243e011945bc1097877909e725ab650bbdd457589" gracePeriod=600 Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.304794 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="config-reloader" containerID="cri-o://0c4f7feff9667adcf34feb5ec56a6ee4ca861aaa9269eb8bb7c66bf6e117df5b" gracePeriod=600 Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.980638 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.990900 4857 generic.go:334] "Generic (PLEG): container finished" podID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" containerID="134f2bba3e91ccb269c800ef4cdd79e63cf2a734297141290056902d1d9e4c57" exitCode=137 Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.990984 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.991089 4857 scope.go:117] "RemoveContainer" containerID="134f2bba3e91ccb269c800ef4cdd79e63cf2a734297141290056902d1d9e4c57" Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.994923 4857 generic.go:334] "Generic (PLEG): container finished" podID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerID="dc12ccdb703732ffdb5d672243e011945bc1097877909e725ab650bbdd457589" exitCode=0 Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.994950 4857 generic.go:334] "Generic (PLEG): container finished" podID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerID="0c4f7feff9667adcf34feb5ec56a6ee4ca861aaa9269eb8bb7c66bf6e117df5b" exitCode=0 Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.994957 4857 generic.go:334] "Generic (PLEG): container finished" podID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerID="d4e54e17db288345bb6a6d7443248544c5e620caa42354563068faf1ae1f4ebd" exitCode=0 Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.995007 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"28a6d05a-72fa-4f72-a3d1-d6891412c04f","Type":"ContainerDied","Data":"dc12ccdb703732ffdb5d672243e011945bc1097877909e725ab650bbdd457589"} Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.995231 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"28a6d05a-72fa-4f72-a3d1-d6891412c04f","Type":"ContainerDied","Data":"0c4f7feff9667adcf34feb5ec56a6ee4ca861aaa9269eb8bb7c66bf6e117df5b"} Dec 01 22:32:56 crc kubenswrapper[4857]: I1201 22:32:56.995254 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"28a6d05a-72fa-4f72-a3d1-d6891412c04f","Type":"ContainerDied","Data":"d4e54e17db288345bb6a6d7443248544c5e620caa42354563068faf1ae1f4ebd"} Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.002242 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49nzh\" (UniqueName: \"kubernetes.io/projected/2f50f03d-e84a-4c7e-a646-047bcfd3046f-kube-api-access-49nzh\") pod \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.002284 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2f50f03d-e84a-4c7e-a646-047bcfd3046f-openstack-config\") pod \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.002325 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f50f03d-e84a-4c7e-a646-047bcfd3046f-combined-ca-bundle\") pod \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.002479 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2f50f03d-e84a-4c7e-a646-047bcfd3046f-openstack-config-secret\") pod \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\" (UID: \"2f50f03d-e84a-4c7e-a646-047bcfd3046f\") " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.012419 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f50f03d-e84a-4c7e-a646-047bcfd3046f-kube-api-access-49nzh" (OuterVolumeSpecName: "kube-api-access-49nzh") pod "2f50f03d-e84a-4c7e-a646-047bcfd3046f" (UID: "2f50f03d-e84a-4c7e-a646-047bcfd3046f"). InnerVolumeSpecName "kube-api-access-49nzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.029864 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f50f03d-e84a-4c7e-a646-047bcfd3046f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "2f50f03d-e84a-4c7e-a646-047bcfd3046f" (UID: "2f50f03d-e84a-4c7e-a646-047bcfd3046f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.052539 4857 scope.go:117] "RemoveContainer" containerID="134f2bba3e91ccb269c800ef4cdd79e63cf2a734297141290056902d1d9e4c57" Dec 01 22:32:57 crc kubenswrapper[4857]: E1201 22:32:57.054461 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"134f2bba3e91ccb269c800ef4cdd79e63cf2a734297141290056902d1d9e4c57\": container with ID starting with 134f2bba3e91ccb269c800ef4cdd79e63cf2a734297141290056902d1d9e4c57 not found: ID does not exist" containerID="134f2bba3e91ccb269c800ef4cdd79e63cf2a734297141290056902d1d9e4c57" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.054488 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"134f2bba3e91ccb269c800ef4cdd79e63cf2a734297141290056902d1d9e4c57"} err="failed to get container status \"134f2bba3e91ccb269c800ef4cdd79e63cf2a734297141290056902d1d9e4c57\": rpc error: code = NotFound desc = could not find container \"134f2bba3e91ccb269c800ef4cdd79e63cf2a734297141290056902d1d9e4c57\": container with ID starting with 134f2bba3e91ccb269c800ef4cdd79e63cf2a734297141290056902d1d9e4c57 not found: ID does not exist" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.065825 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f50f03d-e84a-4c7e-a646-047bcfd3046f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f50f03d-e84a-4c7e-a646-047bcfd3046f" (UID: "2f50f03d-e84a-4c7e-a646-047bcfd3046f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.080225 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f50f03d-e84a-4c7e-a646-047bcfd3046f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "2f50f03d-e84a-4c7e-a646-047bcfd3046f" (UID: "2f50f03d-e84a-4c7e-a646-047bcfd3046f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.105271 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49nzh\" (UniqueName: \"kubernetes.io/projected/2f50f03d-e84a-4c7e-a646-047bcfd3046f-kube-api-access-49nzh\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.105297 4857 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2f50f03d-e84a-4c7e-a646-047bcfd3046f-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.105308 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f50f03d-e84a-4c7e-a646-047bcfd3046f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.105317 4857 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2f50f03d-e84a-4c7e-a646-047bcfd3046f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.309666 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.311980 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" podUID="8b715e70-6964-4dd3-9c88-67d41fd7ad7a" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.410503 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92ngc\" (UniqueName: \"kubernetes.io/projected/28a6d05a-72fa-4f72-a3d1-d6891412c04f-kube-api-access-92ngc\") pod \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.410642 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-thanos-prometheus-http-client-file\") pod \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.410733 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-web-config\") pod \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.410804 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-config\") pod \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.410852 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/28a6d05a-72fa-4f72-a3d1-d6891412c04f-prometheus-metric-storage-rulefiles-0\") pod \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.410886 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.410963 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/28a6d05a-72fa-4f72-a3d1-d6891412c04f-config-out\") pod \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.411005 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/28a6d05a-72fa-4f72-a3d1-d6891412c04f-tls-assets\") pod \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\" (UID: \"28a6d05a-72fa-4f72-a3d1-d6891412c04f\") " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.411905 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28a6d05a-72fa-4f72-a3d1-d6891412c04f-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "28a6d05a-72fa-4f72-a3d1-d6891412c04f" (UID: "28a6d05a-72fa-4f72-a3d1-d6891412c04f"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.416214 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28a6d05a-72fa-4f72-a3d1-d6891412c04f-config-out" (OuterVolumeSpecName: "config-out") pod "28a6d05a-72fa-4f72-a3d1-d6891412c04f" (UID: "28a6d05a-72fa-4f72-a3d1-d6891412c04f"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.416242 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-config" (OuterVolumeSpecName: "config") pod "28a6d05a-72fa-4f72-a3d1-d6891412c04f" (UID: "28a6d05a-72fa-4f72-a3d1-d6891412c04f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.416973 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28a6d05a-72fa-4f72-a3d1-d6891412c04f-kube-api-access-92ngc" (OuterVolumeSpecName: "kube-api-access-92ngc") pod "28a6d05a-72fa-4f72-a3d1-d6891412c04f" (UID: "28a6d05a-72fa-4f72-a3d1-d6891412c04f"). InnerVolumeSpecName "kube-api-access-92ngc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.422536 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28a6d05a-72fa-4f72-a3d1-d6891412c04f-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "28a6d05a-72fa-4f72-a3d1-d6891412c04f" (UID: "28a6d05a-72fa-4f72-a3d1-d6891412c04f"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.423208 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "28a6d05a-72fa-4f72-a3d1-d6891412c04f" (UID: "28a6d05a-72fa-4f72-a3d1-d6891412c04f"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.423805 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "28a6d05a-72fa-4f72-a3d1-d6891412c04f" (UID: "28a6d05a-72fa-4f72-a3d1-d6891412c04f"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.442501 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-web-config" (OuterVolumeSpecName: "web-config") pod "28a6d05a-72fa-4f72-a3d1-d6891412c04f" (UID: "28a6d05a-72fa-4f72-a3d1-d6891412c04f"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.513761 4857 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/28a6d05a-72fa-4f72-a3d1-d6891412c04f-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.513825 4857 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.513837 4857 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/28a6d05a-72fa-4f72-a3d1-d6891412c04f-config-out\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.513847 4857 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/28a6d05a-72fa-4f72-a3d1-d6891412c04f-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.513860 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92ngc\" (UniqueName: \"kubernetes.io/projected/28a6d05a-72fa-4f72-a3d1-d6891412c04f-kube-api-access-92ngc\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.513870 4857 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.513881 4857 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-web-config\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.513889 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/28a6d05a-72fa-4f72-a3d1-d6891412c04f-config\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.542506 4857 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.615583 4857 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 01 22:32:57 crc kubenswrapper[4857]: I1201 22:32:57.846899 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f50f03d-e84a-4c7e-a646-047bcfd3046f" path="/var/lib/kubelet/pods/2f50f03d-e84a-4c7e-a646-047bcfd3046f/volumes" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.012279 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"28a6d05a-72fa-4f72-a3d1-d6891412c04f","Type":"ContainerDied","Data":"558a5f392b336c53e5059f355db55af8e3ef04312ef00e0ba2474404bc9c3f21"} Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.012356 4857 scope.go:117] "RemoveContainer" containerID="dc12ccdb703732ffdb5d672243e011945bc1097877909e725ab650bbdd457589" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.012407 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.049259 4857 scope.go:117] "RemoveContainer" containerID="0c4f7feff9667adcf34feb5ec56a6ee4ca861aaa9269eb8bb7c66bf6e117df5b" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.053250 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.071869 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.082820 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:32:58 crc kubenswrapper[4857]: E1201 22:32:58.083597 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="config-reloader" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.083688 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="config-reloader" Dec 01 22:32:58 crc kubenswrapper[4857]: E1201 22:32:58.083773 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="prometheus" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.083838 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="prometheus" Dec 01 22:32:58 crc kubenswrapper[4857]: E1201 22:32:58.083903 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="thanos-sidecar" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.083964 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="thanos-sidecar" Dec 01 22:32:58 crc kubenswrapper[4857]: E1201 22:32:58.084067 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="init-config-reloader" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.084135 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="init-config-reloader" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.084481 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="prometheus" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.084612 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="thanos-sidecar" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.084694 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" containerName="config-reloader" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.087108 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.094438 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-whwdt" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.095062 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.095374 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.095614 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.097154 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.098395 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.099080 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.099347 4857 scope.go:117] "RemoveContainer" containerID="d4e54e17db288345bb6a6d7443248544c5e620caa42354563068faf1ae1f4ebd" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.102383 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.127637 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-config\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.127709 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/753ef88e-a1c8-4412-9237-e5027d460eb8-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.127898 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/753ef88e-a1c8-4412-9237-e5027d460eb8-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.127925 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.127957 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.128002 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.128099 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.128132 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.128171 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.128205 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/753ef88e-a1c8-4412-9237-e5027d460eb8-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.128362 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prs5v\" (UniqueName: \"kubernetes.io/projected/753ef88e-a1c8-4412-9237-e5027d460eb8-kube-api-access-prs5v\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.133664 4857 scope.go:117] "RemoveContainer" containerID="7eb7848446e6cecd845451a88545310fcff1786bb62581ca40a7411832c137c1" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.230289 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/753ef88e-a1c8-4412-9237-e5027d460eb8-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.230382 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prs5v\" (UniqueName: \"kubernetes.io/projected/753ef88e-a1c8-4412-9237-e5027d460eb8-kube-api-access-prs5v\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.230425 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-config\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.230449 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/753ef88e-a1c8-4412-9237-e5027d460eb8-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.230516 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/753ef88e-a1c8-4412-9237-e5027d460eb8-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.230534 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.230556 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.230580 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.230614 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.230636 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.230660 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.231090 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.231628 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/753ef88e-a1c8-4412-9237-e5027d460eb8-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.234555 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-config\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.234903 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.235876 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.236304 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/753ef88e-a1c8-4412-9237-e5027d460eb8-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.240641 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.241005 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.246615 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.248317 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/753ef88e-a1c8-4412-9237-e5027d460eb8-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.251837 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prs5v\" (UniqueName: \"kubernetes.io/projected/753ef88e-a1c8-4412-9237-e5027d460eb8-kube-api-access-prs5v\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.284676 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.415807 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:32:58 crc kubenswrapper[4857]: I1201 22:32:58.678379 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:32:58 crc kubenswrapper[4857]: W1201 22:32:58.684442 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod753ef88e_a1c8_4412_9237_e5027d460eb8.slice/crio-994c089840b7d8e5898c52c8110e8f3a064734a2029a72d377304a2623acdb1a WatchSource:0}: Error finding container 994c089840b7d8e5898c52c8110e8f3a064734a2029a72d377304a2623acdb1a: Status 404 returned error can't find the container with id 994c089840b7d8e5898c52c8110e8f3a064734a2029a72d377304a2623acdb1a Dec 01 22:32:59 crc kubenswrapper[4857]: I1201 22:32:59.025891 4857 generic.go:334] "Generic (PLEG): container finished" podID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerID="7f0cbeb943e426c39f61ced53e6f1266fc9a5661fa554d4a48415740ecd5d669" exitCode=0 Dec 01 22:32:59 crc kubenswrapper[4857]: I1201 22:32:59.025989 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d2d444dd-0c4b-48ed-927f-9e12b80b9c26","Type":"ContainerDied","Data":"7f0cbeb943e426c39f61ced53e6f1266fc9a5661fa554d4a48415740ecd5d669"} Dec 01 22:32:59 crc kubenswrapper[4857]: I1201 22:32:59.027644 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"753ef88e-a1c8-4412-9237-e5027d460eb8","Type":"ContainerStarted","Data":"994c089840b7d8e5898c52c8110e8f3a064734a2029a72d377304a2623acdb1a"} Dec 01 22:32:59 crc kubenswrapper[4857]: I1201 22:32:59.861915 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28a6d05a-72fa-4f72-a3d1-d6891412c04f" path="/var/lib/kubelet/pods/28a6d05a-72fa-4f72-a3d1-d6891412c04f/volumes" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.044195 4857 generic.go:334] "Generic (PLEG): container finished" podID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerID="5144a43c6f9c4eada7f058dcd0cbb90d3df519122f7d62fafebdd27443472174" exitCode=0 Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.044303 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d2d444dd-0c4b-48ed-927f-9e12b80b9c26","Type":"ContainerDied","Data":"5144a43c6f9c4eada7f058dcd0cbb90d3df519122f7d62fafebdd27443472174"} Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.044573 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d2d444dd-0c4b-48ed-927f-9e12b80b9c26","Type":"ContainerDied","Data":"4c8b4d2fb1d2f5ed4bc544eb5eca7c6f8ce41c5b8d35039cf8e75db4672d810b"} Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.044597 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c8b4d2fb1d2f5ed4bc544eb5eca7c6f8ce41c5b8d35039cf8e75db4672d810b" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.092129 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.272084 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8xdw\" (UniqueName: \"kubernetes.io/projected/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-kube-api-access-r8xdw\") pod \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.272243 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-public-tls-certs\") pod \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.272296 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-internal-tls-certs\") pod \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.272425 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-scripts\") pod \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.272473 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-config-data\") pod \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.272530 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-combined-ca-bundle\") pod \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\" (UID: \"d2d444dd-0c4b-48ed-927f-9e12b80b9c26\") " Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.279825 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-kube-api-access-r8xdw" (OuterVolumeSpecName: "kube-api-access-r8xdw") pod "d2d444dd-0c4b-48ed-927f-9e12b80b9c26" (UID: "d2d444dd-0c4b-48ed-927f-9e12b80b9c26"). InnerVolumeSpecName "kube-api-access-r8xdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.280450 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-scripts" (OuterVolumeSpecName: "scripts") pod "d2d444dd-0c4b-48ed-927f-9e12b80b9c26" (UID: "d2d444dd-0c4b-48ed-927f-9e12b80b9c26"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.334651 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d2d444dd-0c4b-48ed-927f-9e12b80b9c26" (UID: "d2d444dd-0c4b-48ed-927f-9e12b80b9c26"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.362176 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d2d444dd-0c4b-48ed-927f-9e12b80b9c26" (UID: "d2d444dd-0c4b-48ed-927f-9e12b80b9c26"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.374889 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8xdw\" (UniqueName: \"kubernetes.io/projected/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-kube-api-access-r8xdw\") on node \"crc\" DevicePath \"\"" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.374937 4857 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.374947 4857 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.374959 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.398666 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-config-data" (OuterVolumeSpecName: "config-data") pod "d2d444dd-0c4b-48ed-927f-9e12b80b9c26" (UID: "d2d444dd-0c4b-48ed-927f-9e12b80b9c26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.410824 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2d444dd-0c4b-48ed-927f-9e12b80b9c26" (UID: "d2d444dd-0c4b-48ed-927f-9e12b80b9c26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.477439 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:33:00 crc kubenswrapper[4857]: I1201 22:33:00.477474 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d444dd-0c4b-48ed-927f-9e12b80b9c26-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.135133 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.179816 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.197464 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.208995 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 01 22:33:01 crc kubenswrapper[4857]: E1201 22:33:01.209651 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-evaluator" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.209716 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-evaluator" Dec 01 22:33:01 crc kubenswrapper[4857]: E1201 22:33:01.209798 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-listener" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.209846 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-listener" Dec 01 22:33:01 crc kubenswrapper[4857]: E1201 22:33:01.209898 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-notifier" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.209942 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-notifier" Dec 01 22:33:01 crc kubenswrapper[4857]: E1201 22:33:01.210017 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-api" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.210200 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-api" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.210461 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-evaluator" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.210524 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-listener" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.210594 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-api" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.210641 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" containerName="aodh-notifier" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.212977 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.218766 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.218931 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.219123 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.218891 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-j6fzf" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.218923 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.218982 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.232117 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-scripts\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.232174 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-combined-ca-bundle\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.232239 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-public-tls-certs\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.232291 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7zrs\" (UniqueName: \"kubernetes.io/projected/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-kube-api-access-b7zrs\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.232314 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-internal-tls-certs\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.232349 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-config-data\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.334366 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-scripts\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.334466 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-combined-ca-bundle\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.334564 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-public-tls-certs\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.334674 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7zrs\" (UniqueName: \"kubernetes.io/projected/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-kube-api-access-b7zrs\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.334712 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-internal-tls-certs\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.334784 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-config-data\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.338440 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-scripts\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.339333 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-public-tls-certs\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.339866 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-internal-tls-certs\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.340747 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-combined-ca-bundle\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.340813 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-config-data\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:01 crc kubenswrapper[4857]: I1201 22:33:01.351775 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7zrs\" (UniqueName: \"kubernetes.io/projected/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-kube-api-access-b7zrs\") pod \"aodh-0\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " pod="openstack/aodh-0" Dec 01 22:33:02 crc kubenswrapper[4857]: I1201 22:33:01.543881 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:33:02 crc kubenswrapper[4857]: I1201 22:33:01.860665 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2d444dd-0c4b-48ed-927f-9e12b80b9c26" path="/var/lib/kubelet/pods/d2d444dd-0c4b-48ed-927f-9e12b80b9c26/volumes" Dec 01 22:33:02 crc kubenswrapper[4857]: I1201 22:33:02.432568 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 22:33:02 crc kubenswrapper[4857]: W1201 22:33:02.438055 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod899ce6c7_16e9_49d0_bdf9_5aeceb6ddebd.slice/crio-faa8bd9c2632c5af77e7d15b46bc0a5ca78da8bd78101ddcf86779da477f7f44 WatchSource:0}: Error finding container faa8bd9c2632c5af77e7d15b46bc0a5ca78da8bd78101ddcf86779da477f7f44: Status 404 returned error can't find the container with id faa8bd9c2632c5af77e7d15b46bc0a5ca78da8bd78101ddcf86779da477f7f44 Dec 01 22:33:03 crc kubenswrapper[4857]: I1201 22:33:03.169364 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd","Type":"ContainerStarted","Data":"faa8bd9c2632c5af77e7d15b46bc0a5ca78da8bd78101ddcf86779da477f7f44"} Dec 01 22:33:03 crc kubenswrapper[4857]: I1201 22:33:03.174719 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"753ef88e-a1c8-4412-9237-e5027d460eb8","Type":"ContainerStarted","Data":"4abf17b05fc4de399750addfbbc24c5ac5a1df6beb4e177865a27f7b4704abc7"} Dec 01 22:33:04 crc kubenswrapper[4857]: I1201 22:33:04.187916 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd","Type":"ContainerStarted","Data":"fc80ee16bf37bb1d8989461b54d93f34d7dd052d2d5b7d5e59fd47ead6c4af8a"} Dec 01 22:33:04 crc kubenswrapper[4857]: I1201 22:33:04.188483 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd","Type":"ContainerStarted","Data":"9edebc054f83fd8d57bbcc768756836db2bbbd12e496f9b248732a24ec49a41a"} Dec 01 22:33:05 crc kubenswrapper[4857]: I1201 22:33:05.200293 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd","Type":"ContainerStarted","Data":"43c74588a23e5156a62edee6cdb450e340eaa287a560f3fae550cdbe09941108"} Dec 01 22:33:06 crc kubenswrapper[4857]: I1201 22:33:06.212668 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd","Type":"ContainerStarted","Data":"e3dcf6edfbbc244c1a24657f03709f63b5c8e30e1e8a0c73127a254f599d02fb"} Dec 01 22:33:12 crc kubenswrapper[4857]: I1201 22:33:12.282146 4857 generic.go:334] "Generic (PLEG): container finished" podID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerID="4abf17b05fc4de399750addfbbc24c5ac5a1df6beb4e177865a27f7b4704abc7" exitCode=0 Dec 01 22:33:12 crc kubenswrapper[4857]: I1201 22:33:12.282248 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"753ef88e-a1c8-4412-9237-e5027d460eb8","Type":"ContainerDied","Data":"4abf17b05fc4de399750addfbbc24c5ac5a1df6beb4e177865a27f7b4704abc7"} Dec 01 22:33:12 crc kubenswrapper[4857]: I1201 22:33:12.348754 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=7.959991614 podStartE2EDuration="11.348727478s" podCreationTimestamp="2025-12-01 22:33:01 +0000 UTC" firstStartedPulling="2025-12-01 22:33:02.441574415 +0000 UTC m=+3500.931636762" lastFinishedPulling="2025-12-01 22:33:05.830310279 +0000 UTC m=+3504.320372626" observedRunningTime="2025-12-01 22:33:06.244426073 +0000 UTC m=+3504.734488400" watchObservedRunningTime="2025-12-01 22:33:12.348727478 +0000 UTC m=+3510.838789835" Dec 01 22:33:13 crc kubenswrapper[4857]: I1201 22:33:13.301675 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"753ef88e-a1c8-4412-9237-e5027d460eb8","Type":"ContainerStarted","Data":"f18db5f3196d1483de12cd229eeff07ef4b649b08cbe7c83f540b8315a9a2498"} Dec 01 22:33:18 crc kubenswrapper[4857]: I1201 22:33:18.374962 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"753ef88e-a1c8-4412-9237-e5027d460eb8","Type":"ContainerStarted","Data":"60f0f29e1f7b0ed3dd4c2e3b1b05877dd42f5391633e38acfefeb9c339af019b"} Dec 01 22:33:18 crc kubenswrapper[4857]: I1201 22:33:18.375621 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"753ef88e-a1c8-4412-9237-e5027d460eb8","Type":"ContainerStarted","Data":"cd6aa665731181ded790e300cde334f18705b5152a0b57d1f2fb064e738a3b7c"} Dec 01 22:33:18 crc kubenswrapper[4857]: I1201 22:33:18.416209 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 01 22:33:18 crc kubenswrapper[4857]: I1201 22:33:18.423656 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=20.423625181 podStartE2EDuration="20.423625181s" podCreationTimestamp="2025-12-01 22:32:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 22:33:18.404575514 +0000 UTC m=+3516.894637871" watchObservedRunningTime="2025-12-01 22:33:18.423625181 +0000 UTC m=+3516.913687528" Dec 01 22:33:28 crc kubenswrapper[4857]: I1201 22:33:28.416890 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 01 22:33:28 crc kubenswrapper[4857]: I1201 22:33:28.427893 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 01 22:33:28 crc kubenswrapper[4857]: I1201 22:33:28.513170 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 01 22:34:23 crc kubenswrapper[4857]: I1201 22:34:23.892811 4857 scope.go:117] "RemoveContainer" containerID="996fefaac23ac63ff0afd65bea71df4c08d41a78268be04474cc652807258715" Dec 01 22:34:23 crc kubenswrapper[4857]: I1201 22:34:23.937004 4857 scope.go:117] "RemoveContainer" containerID="4702163fa4c65d550ec70cb4b1c15e4c93137d835ae245ef328119d26963f795" Dec 01 22:34:23 crc kubenswrapper[4857]: I1201 22:34:23.962388 4857 scope.go:117] "RemoveContainer" containerID="d39378bf79c21b3adfa86e270c6cd68b9434e065b3e0a66169d33e26935bd602" Dec 01 22:34:23 crc kubenswrapper[4857]: I1201 22:34:23.990402 4857 scope.go:117] "RemoveContainer" containerID="06ce657f9604cee3e2f96b8b52ced69d57105d8a456835960424448358c665fc" Dec 01 22:34:27 crc kubenswrapper[4857]: I1201 22:34:27.806583 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:34:27 crc kubenswrapper[4857]: I1201 22:34:27.807291 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:34:57 crc kubenswrapper[4857]: I1201 22:34:57.615807 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f765fc7d6-x797d_ff230e58-1711-4638-89d6-daf91fb3bfc1/manager/0.log" Dec 01 22:34:57 crc kubenswrapper[4857]: I1201 22:34:57.806858 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:34:57 crc kubenswrapper[4857]: I1201 22:34:57.806962 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:35:00 crc kubenswrapper[4857]: I1201 22:35:00.249488 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:35:00 crc kubenswrapper[4857]: I1201 22:35:00.250034 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="prometheus" containerID="cri-o://f18db5f3196d1483de12cd229eeff07ef4b649b08cbe7c83f540b8315a9a2498" gracePeriod=600 Dec 01 22:35:00 crc kubenswrapper[4857]: I1201 22:35:00.250672 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="thanos-sidecar" containerID="cri-o://60f0f29e1f7b0ed3dd4c2e3b1b05877dd42f5391633e38acfefeb9c339af019b" gracePeriod=600 Dec 01 22:35:00 crc kubenswrapper[4857]: I1201 22:35:00.250960 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="config-reloader" containerID="cri-o://cd6aa665731181ded790e300cde334f18705b5152a0b57d1f2fb064e738a3b7c" gracePeriod=600 Dec 01 22:35:00 crc kubenswrapper[4857]: I1201 22:35:00.781930 4857 generic.go:334] "Generic (PLEG): container finished" podID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerID="60f0f29e1f7b0ed3dd4c2e3b1b05877dd42f5391633e38acfefeb9c339af019b" exitCode=0 Dec 01 22:35:00 crc kubenswrapper[4857]: I1201 22:35:00.782328 4857 generic.go:334] "Generic (PLEG): container finished" podID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerID="f18db5f3196d1483de12cd229eeff07ef4b649b08cbe7c83f540b8315a9a2498" exitCode=0 Dec 01 22:35:00 crc kubenswrapper[4857]: I1201 22:35:00.781977 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"753ef88e-a1c8-4412-9237-e5027d460eb8","Type":"ContainerDied","Data":"60f0f29e1f7b0ed3dd4c2e3b1b05877dd42f5391633e38acfefeb9c339af019b"} Dec 01 22:35:00 crc kubenswrapper[4857]: I1201 22:35:00.782393 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"753ef88e-a1c8-4412-9237-e5027d460eb8","Type":"ContainerDied","Data":"f18db5f3196d1483de12cd229eeff07ef4b649b08cbe7c83f540b8315a9a2498"} Dec 01 22:35:01 crc kubenswrapper[4857]: I1201 22:35:01.814009 4857 generic.go:334] "Generic (PLEG): container finished" podID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerID="cd6aa665731181ded790e300cde334f18705b5152a0b57d1f2fb064e738a3b7c" exitCode=0 Dec 01 22:35:01 crc kubenswrapper[4857]: I1201 22:35:01.814090 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"753ef88e-a1c8-4412-9237-e5027d460eb8","Type":"ContainerDied","Data":"cd6aa665731181ded790e300cde334f18705b5152a0b57d1f2fb064e738a3b7c"} Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.071892 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.249613 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-thanos-prometheus-http-client-file\") pod \"753ef88e-a1c8-4412-9237-e5027d460eb8\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.249659 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"753ef88e-a1c8-4412-9237-e5027d460eb8\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.249760 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"753ef88e-a1c8-4412-9237-e5027d460eb8\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.249918 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-secret-combined-ca-bundle\") pod \"753ef88e-a1c8-4412-9237-e5027d460eb8\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.249992 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/753ef88e-a1c8-4412-9237-e5027d460eb8-prometheus-metric-storage-rulefiles-0\") pod \"753ef88e-a1c8-4412-9237-e5027d460eb8\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.250026 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/753ef88e-a1c8-4412-9237-e5027d460eb8-config-out\") pod \"753ef88e-a1c8-4412-9237-e5027d460eb8\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.250085 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/753ef88e-a1c8-4412-9237-e5027d460eb8-tls-assets\") pod \"753ef88e-a1c8-4412-9237-e5027d460eb8\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.250106 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prs5v\" (UniqueName: \"kubernetes.io/projected/753ef88e-a1c8-4412-9237-e5027d460eb8-kube-api-access-prs5v\") pod \"753ef88e-a1c8-4412-9237-e5027d460eb8\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.250174 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-config\") pod \"753ef88e-a1c8-4412-9237-e5027d460eb8\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.250303 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config\") pod \"753ef88e-a1c8-4412-9237-e5027d460eb8\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.250342 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"753ef88e-a1c8-4412-9237-e5027d460eb8\" (UID: \"753ef88e-a1c8-4412-9237-e5027d460eb8\") " Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.251722 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/753ef88e-a1c8-4412-9237-e5027d460eb8-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "753ef88e-a1c8-4412-9237-e5027d460eb8" (UID: "753ef88e-a1c8-4412-9237-e5027d460eb8"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.257535 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "753ef88e-a1c8-4412-9237-e5027d460eb8" (UID: "753ef88e-a1c8-4412-9237-e5027d460eb8"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.258320 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-config" (OuterVolumeSpecName: "config") pod "753ef88e-a1c8-4412-9237-e5027d460eb8" (UID: "753ef88e-a1c8-4412-9237-e5027d460eb8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.261239 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/753ef88e-a1c8-4412-9237-e5027d460eb8-kube-api-access-prs5v" (OuterVolumeSpecName: "kube-api-access-prs5v") pod "753ef88e-a1c8-4412-9237-e5027d460eb8" (UID: "753ef88e-a1c8-4412-9237-e5027d460eb8"). InnerVolumeSpecName "kube-api-access-prs5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.261275 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/753ef88e-a1c8-4412-9237-e5027d460eb8-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "753ef88e-a1c8-4412-9237-e5027d460eb8" (UID: "753ef88e-a1c8-4412-9237-e5027d460eb8"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.261283 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/753ef88e-a1c8-4412-9237-e5027d460eb8-config-out" (OuterVolumeSpecName: "config-out") pod "753ef88e-a1c8-4412-9237-e5027d460eb8" (UID: "753ef88e-a1c8-4412-9237-e5027d460eb8"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.262126 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "753ef88e-a1c8-4412-9237-e5027d460eb8" (UID: "753ef88e-a1c8-4412-9237-e5027d460eb8"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.262532 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "753ef88e-a1c8-4412-9237-e5027d460eb8" (UID: "753ef88e-a1c8-4412-9237-e5027d460eb8"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.264221 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "753ef88e-a1c8-4412-9237-e5027d460eb8" (UID: "753ef88e-a1c8-4412-9237-e5027d460eb8"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.266158 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "753ef88e-a1c8-4412-9237-e5027d460eb8" (UID: "753ef88e-a1c8-4412-9237-e5027d460eb8"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.352856 4857 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.352944 4857 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.352969 4857 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/753ef88e-a1c8-4412-9237-e5027d460eb8-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.352992 4857 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/753ef88e-a1c8-4412-9237-e5027d460eb8-config-out\") on node \"crc\" DevicePath \"\"" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.353017 4857 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/753ef88e-a1c8-4412-9237-e5027d460eb8-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.353090 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prs5v\" (UniqueName: \"kubernetes.io/projected/753ef88e-a1c8-4412-9237-e5027d460eb8-kube-api-access-prs5v\") on node \"crc\" DevicePath \"\"" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.353111 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-config\") on node \"crc\" DevicePath \"\"" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.353150 4857 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.353171 4857 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.353193 4857 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.384612 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config" (OuterVolumeSpecName: "web-config") pod "753ef88e-a1c8-4412-9237-e5027d460eb8" (UID: "753ef88e-a1c8-4412-9237-e5027d460eb8"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.397432 4857 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.455742 4857 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.456212 4857 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/753ef88e-a1c8-4412-9237-e5027d460eb8-web-config\") on node \"crc\" DevicePath \"\"" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.832440 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"753ef88e-a1c8-4412-9237-e5027d460eb8","Type":"ContainerDied","Data":"994c089840b7d8e5898c52c8110e8f3a064734a2029a72d377304a2623acdb1a"} Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.832530 4857 scope.go:117] "RemoveContainer" containerID="60f0f29e1f7b0ed3dd4c2e3b1b05877dd42f5391633e38acfefeb9c339af019b" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.832621 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.873209 4857 scope.go:117] "RemoveContainer" containerID="cd6aa665731181ded790e300cde334f18705b5152a0b57d1f2fb064e738a3b7c" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.904298 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.917995 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.922894 4857 scope.go:117] "RemoveContainer" containerID="f18db5f3196d1483de12cd229eeff07ef4b649b08cbe7c83f540b8315a9a2498" Dec 01 22:35:02 crc kubenswrapper[4857]: I1201 22:35:02.958319 4857 scope.go:117] "RemoveContainer" containerID="4abf17b05fc4de399750addfbbc24c5ac5a1df6beb4e177865a27f7b4704abc7" Dec 01 22:35:03 crc kubenswrapper[4857]: I1201 22:35:03.849876 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" path="/var/lib/kubelet/pods/753ef88e-a1c8-4412-9237-e5027d460eb8/volumes" Dec 01 22:35:03 crc kubenswrapper[4857]: I1201 22:35:03.977353 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:35:03 crc kubenswrapper[4857]: E1201 22:35:03.977914 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="prometheus" Dec 01 22:35:03 crc kubenswrapper[4857]: I1201 22:35:03.977947 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="prometheus" Dec 01 22:35:03 crc kubenswrapper[4857]: E1201 22:35:03.977987 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="thanos-sidecar" Dec 01 22:35:03 crc kubenswrapper[4857]: I1201 22:35:03.978000 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="thanos-sidecar" Dec 01 22:35:03 crc kubenswrapper[4857]: E1201 22:35:03.978024 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="init-config-reloader" Dec 01 22:35:03 crc kubenswrapper[4857]: I1201 22:35:03.978129 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="init-config-reloader" Dec 01 22:35:03 crc kubenswrapper[4857]: E1201 22:35:03.978153 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="config-reloader" Dec 01 22:35:03 crc kubenswrapper[4857]: I1201 22:35:03.978179 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="config-reloader" Dec 01 22:35:03 crc kubenswrapper[4857]: I1201 22:35:03.978429 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="prometheus" Dec 01 22:35:03 crc kubenswrapper[4857]: I1201 22:35:03.978463 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="thanos-sidecar" Dec 01 22:35:03 crc kubenswrapper[4857]: I1201 22:35:03.978476 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="753ef88e-a1c8-4412-9237-e5027d460eb8" containerName="config-reloader" Dec 01 22:35:03 crc kubenswrapper[4857]: I1201 22:35:03.982170 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.001476 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.001539 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.001596 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.001731 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.001958 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.001970 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.002287 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-whwdt" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.019502 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.095929 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/84257e31-0d4a-4b6f-857f-fb396eb2faf7-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.096054 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-config\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.096114 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.096449 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/84257e31-0d4a-4b6f-857f-fb396eb2faf7-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.096482 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.096504 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.096520 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59gtg\" (UniqueName: \"kubernetes.io/projected/84257e31-0d4a-4b6f-857f-fb396eb2faf7-kube-api-access-59gtg\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.096554 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/84257e31-0d4a-4b6f-857f-fb396eb2faf7-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.096572 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/84257e31-0d4a-4b6f-857f-fb396eb2faf7-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.096623 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.096666 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.198648 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/84257e31-0d4a-4b6f-857f-fb396eb2faf7-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.198702 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.198735 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.198757 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59gtg\" (UniqueName: \"kubernetes.io/projected/84257e31-0d4a-4b6f-857f-fb396eb2faf7-kube-api-access-59gtg\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.198798 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/84257e31-0d4a-4b6f-857f-fb396eb2faf7-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.198823 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/84257e31-0d4a-4b6f-857f-fb396eb2faf7-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.198852 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.198890 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.198946 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/84257e31-0d4a-4b6f-857f-fb396eb2faf7-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.199020 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-config\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.199063 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.200397 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/84257e31-0d4a-4b6f-857f-fb396eb2faf7-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.200896 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/84257e31-0d4a-4b6f-857f-fb396eb2faf7-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.205205 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/84257e31-0d4a-4b6f-857f-fb396eb2faf7-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.205529 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.205896 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.206458 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.207261 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.208138 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-config\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.218440 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.218872 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/84257e31-0d4a-4b6f-857f-fb396eb2faf7-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.221190 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59gtg\" (UniqueName: \"kubernetes.io/projected/84257e31-0d4a-4b6f-857f-fb396eb2faf7-kube-api-access-59gtg\") pod \"prometheus-metric-storage-0\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.318847 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.858521 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:35:04 crc kubenswrapper[4857]: I1201 22:35:04.882374 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"84257e31-0d4a-4b6f-857f-fb396eb2faf7","Type":"ContainerStarted","Data":"34307c2f1fdecd7c2e1ed08ca099299af155bb57b719c9b0eb04f9f95ef4fecb"} Dec 01 22:35:09 crc kubenswrapper[4857]: I1201 22:35:09.958705 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"84257e31-0d4a-4b6f-857f-fb396eb2faf7","Type":"ContainerStarted","Data":"c3b46f9f80b298dd413728db6458bec4c1dc9e7e80d89d5095ff753ab0177ea8"} Dec 01 22:35:19 crc kubenswrapper[4857]: I1201 22:35:19.079196 4857 generic.go:334] "Generic (PLEG): container finished" podID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerID="c3b46f9f80b298dd413728db6458bec4c1dc9e7e80d89d5095ff753ab0177ea8" exitCode=0 Dec 01 22:35:19 crc kubenswrapper[4857]: I1201 22:35:19.079350 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"84257e31-0d4a-4b6f-857f-fb396eb2faf7","Type":"ContainerDied","Data":"c3b46f9f80b298dd413728db6458bec4c1dc9e7e80d89d5095ff753ab0177ea8"} Dec 01 22:35:20 crc kubenswrapper[4857]: I1201 22:35:20.098573 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"84257e31-0d4a-4b6f-857f-fb396eb2faf7","Type":"ContainerStarted","Data":"508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7"} Dec 01 22:35:25 crc kubenswrapper[4857]: I1201 22:35:25.167202 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"84257e31-0d4a-4b6f-857f-fb396eb2faf7","Type":"ContainerStarted","Data":"cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26"} Dec 01 22:35:25 crc kubenswrapper[4857]: I1201 22:35:25.168071 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"84257e31-0d4a-4b6f-857f-fb396eb2faf7","Type":"ContainerStarted","Data":"eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d"} Dec 01 22:35:25 crc kubenswrapper[4857]: I1201 22:35:25.210295 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=22.210269618 podStartE2EDuration="22.210269618s" podCreationTimestamp="2025-12-01 22:35:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 22:35:25.207230665 +0000 UTC m=+3643.697292992" watchObservedRunningTime="2025-12-01 22:35:25.210269618 +0000 UTC m=+3643.700331965" Dec 01 22:35:27 crc kubenswrapper[4857]: I1201 22:35:27.806689 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:35:27 crc kubenswrapper[4857]: I1201 22:35:27.807118 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:35:27 crc kubenswrapper[4857]: I1201 22:35:27.807186 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 22:35:27 crc kubenswrapper[4857]: I1201 22:35:27.808326 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7018e5ef4fa6765cca44e184b68c5e2a16e99c3def48589b77a1494d6ddc3ee9"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 22:35:27 crc kubenswrapper[4857]: I1201 22:35:27.808428 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://7018e5ef4fa6765cca44e184b68c5e2a16e99c3def48589b77a1494d6ddc3ee9" gracePeriod=600 Dec 01 22:35:28 crc kubenswrapper[4857]: I1201 22:35:28.258098 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"7018e5ef4fa6765cca44e184b68c5e2a16e99c3def48589b77a1494d6ddc3ee9"} Dec 01 22:35:28 crc kubenswrapper[4857]: I1201 22:35:28.258175 4857 scope.go:117] "RemoveContainer" containerID="c1a00193d0b406e83e86d3a16d18f491240a46d8c90712011e6e2e0be2f8e170" Dec 01 22:35:28 crc kubenswrapper[4857]: I1201 22:35:28.258031 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="7018e5ef4fa6765cca44e184b68c5e2a16e99c3def48589b77a1494d6ddc3ee9" exitCode=0 Dec 01 22:35:29 crc kubenswrapper[4857]: I1201 22:35:29.270495 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104"} Dec 01 22:35:29 crc kubenswrapper[4857]: I1201 22:35:29.320004 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:34 crc kubenswrapper[4857]: I1201 22:35:34.320607 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:34 crc kubenswrapper[4857]: I1201 22:35:34.332178 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 01 22:35:34 crc kubenswrapper[4857]: I1201 22:35:34.351856 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 01 22:36:59 crc kubenswrapper[4857]: I1201 22:36:59.577941 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f765fc7d6-x797d_ff230e58-1711-4638-89d6-daf91fb3bfc1/manager/0.log" Dec 01 22:37:01 crc kubenswrapper[4857]: I1201 22:37:01.395205 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 22:37:01 crc kubenswrapper[4857]: I1201 22:37:01.396029 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-api" containerID="cri-o://9edebc054f83fd8d57bbcc768756836db2bbbd12e496f9b248732a24ec49a41a" gracePeriod=30 Dec 01 22:37:01 crc kubenswrapper[4857]: I1201 22:37:01.396537 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-listener" containerID="cri-o://e3dcf6edfbbc244c1a24657f03709f63b5c8e30e1e8a0c73127a254f599d02fb" gracePeriod=30 Dec 01 22:37:01 crc kubenswrapper[4857]: I1201 22:37:01.396592 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-notifier" containerID="cri-o://43c74588a23e5156a62edee6cdb450e340eaa287a560f3fae550cdbe09941108" gracePeriod=30 Dec 01 22:37:01 crc kubenswrapper[4857]: I1201 22:37:01.396625 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-evaluator" containerID="cri-o://fc80ee16bf37bb1d8989461b54d93f34d7dd052d2d5b7d5e59fd47ead6c4af8a" gracePeriod=30 Dec 01 22:37:02 crc kubenswrapper[4857]: I1201 22:37:02.583073 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd","Type":"ContainerDied","Data":"43c74588a23e5156a62edee6cdb450e340eaa287a560f3fae550cdbe09941108"} Dec 01 22:37:02 crc kubenswrapper[4857]: I1201 22:37:02.583073 4857 generic.go:334] "Generic (PLEG): container finished" podID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerID="43c74588a23e5156a62edee6cdb450e340eaa287a560f3fae550cdbe09941108" exitCode=0 Dec 01 22:37:02 crc kubenswrapper[4857]: I1201 22:37:02.583364 4857 generic.go:334] "Generic (PLEG): container finished" podID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerID="fc80ee16bf37bb1d8989461b54d93f34d7dd052d2d5b7d5e59fd47ead6c4af8a" exitCode=0 Dec 01 22:37:02 crc kubenswrapper[4857]: I1201 22:37:02.583379 4857 generic.go:334] "Generic (PLEG): container finished" podID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerID="9edebc054f83fd8d57bbcc768756836db2bbbd12e496f9b248732a24ec49a41a" exitCode=0 Dec 01 22:37:02 crc kubenswrapper[4857]: I1201 22:37:02.583393 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd","Type":"ContainerDied","Data":"fc80ee16bf37bb1d8989461b54d93f34d7dd052d2d5b7d5e59fd47ead6c4af8a"} Dec 01 22:37:02 crc kubenswrapper[4857]: I1201 22:37:02.583403 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd","Type":"ContainerDied","Data":"9edebc054f83fd8d57bbcc768756836db2bbbd12e496f9b248732a24ec49a41a"} Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.599107 4857 generic.go:334] "Generic (PLEG): container finished" podID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerID="e3dcf6edfbbc244c1a24657f03709f63b5c8e30e1e8a0c73127a254f599d02fb" exitCode=0 Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.599188 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd","Type":"ContainerDied","Data":"e3dcf6edfbbc244c1a24657f03709f63b5c8e30e1e8a0c73127a254f599d02fb"} Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.705021 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.742871 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-config-data\") pod \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.742991 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7zrs\" (UniqueName: \"kubernetes.io/projected/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-kube-api-access-b7zrs\") pod \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.743179 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-public-tls-certs\") pod \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.743218 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-combined-ca-bundle\") pod \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.743269 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-scripts\") pod \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.743383 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-internal-tls-certs\") pod \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\" (UID: \"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd\") " Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.767722 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-kube-api-access-b7zrs" (OuterVolumeSpecName: "kube-api-access-b7zrs") pod "899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" (UID: "899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd"). InnerVolumeSpecName "kube-api-access-b7zrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.774944 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-scripts" (OuterVolumeSpecName: "scripts") pod "899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" (UID: "899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.829679 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" (UID: "899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.847984 4857 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.848006 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7zrs\" (UniqueName: \"kubernetes.io/projected/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-kube-api-access-b7zrs\") on node \"crc\" DevicePath \"\"" Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.848018 4857 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.852231 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" (UID: "899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.937902 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" (UID: "899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.938213 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-config-data" (OuterVolumeSpecName: "config-data") pod "899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" (UID: "899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.949841 4857 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.949868 4857 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:37:03 crc kubenswrapper[4857]: I1201 22:37:03.949879 4857 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.614085 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd","Type":"ContainerDied","Data":"faa8bd9c2632c5af77e7d15b46bc0a5ca78da8bd78101ddcf86779da477f7f44"} Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.614169 4857 scope.go:117] "RemoveContainer" containerID="e3dcf6edfbbc244c1a24657f03709f63b5c8e30e1e8a0c73127a254f599d02fb" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.614265 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.647877 4857 scope.go:117] "RemoveContainer" containerID="43c74588a23e5156a62edee6cdb450e340eaa287a560f3fae550cdbe09941108" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.676151 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.682184 4857 scope.go:117] "RemoveContainer" containerID="fc80ee16bf37bb1d8989461b54d93f34d7dd052d2d5b7d5e59fd47ead6c4af8a" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.691628 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.709825 4857 scope.go:117] "RemoveContainer" containerID="9edebc054f83fd8d57bbcc768756836db2bbbd12e496f9b248732a24ec49a41a" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.714975 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 01 22:37:04 crc kubenswrapper[4857]: E1201 22:37:04.715545 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-listener" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.715562 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-listener" Dec 01 22:37:04 crc kubenswrapper[4857]: E1201 22:37:04.715578 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-evaluator" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.715584 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-evaluator" Dec 01 22:37:04 crc kubenswrapper[4857]: E1201 22:37:04.715597 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-notifier" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.715602 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-notifier" Dec 01 22:37:04 crc kubenswrapper[4857]: E1201 22:37:04.715618 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-api" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.715624 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-api" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.715853 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-listener" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.715879 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-notifier" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.715888 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-api" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.715905 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" containerName="aodh-evaluator" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.717858 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.720526 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.720942 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.721217 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.721375 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-j6fzf" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.721538 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.723484 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.763723 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-scripts\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.763843 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-config-data\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.763877 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.763923 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-internal-tls-certs\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.764005 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktlzs\" (UniqueName: \"kubernetes.io/projected/e78cc157-06d0-4c5f-8da8-6e1a9591600a-kube-api-access-ktlzs\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.764146 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-public-tls-certs\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.866334 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-scripts\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.866437 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-config-data\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.866461 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.866518 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-internal-tls-certs\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.866577 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktlzs\" (UniqueName: \"kubernetes.io/projected/e78cc157-06d0-4c5f-8da8-6e1a9591600a-kube-api-access-ktlzs\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.867205 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-public-tls-certs\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.871570 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-scripts\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.872250 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-config-data\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.876553 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-internal-tls-certs\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.876571 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-public-tls-certs\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.877189 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78cc157-06d0-4c5f-8da8-6e1a9591600a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:04 crc kubenswrapper[4857]: I1201 22:37:04.886740 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktlzs\" (UniqueName: \"kubernetes.io/projected/e78cc157-06d0-4c5f-8da8-6e1a9591600a-kube-api-access-ktlzs\") pod \"aodh-0\" (UID: \"e78cc157-06d0-4c5f-8da8-6e1a9591600a\") " pod="openstack/aodh-0" Dec 01 22:37:05 crc kubenswrapper[4857]: I1201 22:37:05.039551 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 22:37:05 crc kubenswrapper[4857]: I1201 22:37:05.040130 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-mh9p4"] Dec 01 22:37:05 crc kubenswrapper[4857]: I1201 22:37:05.052896 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-mh9p4"] Dec 01 22:37:05 crc kubenswrapper[4857]: I1201 22:37:05.552381 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 22:37:05 crc kubenswrapper[4857]: I1201 22:37:05.566094 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 22:37:05 crc kubenswrapper[4857]: I1201 22:37:05.630140 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e78cc157-06d0-4c5f-8da8-6e1a9591600a","Type":"ContainerStarted","Data":"7d589af939c395ed021b4d40aa78b18ed6c759970420f1ad81bd07e4d72b66f7"} Dec 01 22:37:05 crc kubenswrapper[4857]: I1201 22:37:05.863535 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd" path="/var/lib/kubelet/pods/899ce6c7-16e9-49d0-bdf9-5aeceb6ddebd/volumes" Dec 01 22:37:05 crc kubenswrapper[4857]: I1201 22:37:05.865424 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e58803a1-45fc-4be0-9f0e-536338384e04" path="/var/lib/kubelet/pods/e58803a1-45fc-4be0-9f0e-536338384e04/volumes" Dec 01 22:37:06 crc kubenswrapper[4857]: I1201 22:37:06.037869 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-487f-account-create-update-vszxx"] Dec 01 22:37:06 crc kubenswrapper[4857]: I1201 22:37:06.049480 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-487f-account-create-update-vszxx"] Dec 01 22:37:06 crc kubenswrapper[4857]: I1201 22:37:06.640128 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e78cc157-06d0-4c5f-8da8-6e1a9591600a","Type":"ContainerStarted","Data":"02fbdb279df1f6a3d3cde75af9f85aac9fc503520b634afbdc465e6988f66698"} Dec 01 22:37:07 crc kubenswrapper[4857]: I1201 22:37:07.652974 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e78cc157-06d0-4c5f-8da8-6e1a9591600a","Type":"ContainerStarted","Data":"4b70f298baa1d96f154ba2f88078d2d3139affb6a551795c2907464210a674a2"} Dec 01 22:37:07 crc kubenswrapper[4857]: I1201 22:37:07.849013 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0589b0a-71fe-47a1-9dd3-981c43ac56f5" path="/var/lib/kubelet/pods/f0589b0a-71fe-47a1-9dd3-981c43ac56f5/volumes" Dec 01 22:37:08 crc kubenswrapper[4857]: I1201 22:37:08.663654 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e78cc157-06d0-4c5f-8da8-6e1a9591600a","Type":"ContainerStarted","Data":"9c0e683e729f02f50be73ae6d232883d89e3b998ac89093f3f86cc42b13ab6d5"} Dec 01 22:37:09 crc kubenswrapper[4857]: I1201 22:37:09.680332 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e78cc157-06d0-4c5f-8da8-6e1a9591600a","Type":"ContainerStarted","Data":"13aa1e8edbb8ae4342831962e690024f898d58d77d4dcb2401d580c09cc8b36c"} Dec 01 22:37:18 crc kubenswrapper[4857]: I1201 22:37:18.045721 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=10.816891774 podStartE2EDuration="14.045651568s" podCreationTimestamp="2025-12-01 22:37:04 +0000 UTC" firstStartedPulling="2025-12-01 22:37:05.565817928 +0000 UTC m=+3744.055880245" lastFinishedPulling="2025-12-01 22:37:08.794577672 +0000 UTC m=+3747.284640039" observedRunningTime="2025-12-01 22:37:09.714099948 +0000 UTC m=+3748.204162305" watchObservedRunningTime="2025-12-01 22:37:18.045651568 +0000 UTC m=+3756.535713895" Dec 01 22:37:18 crc kubenswrapper[4857]: I1201 22:37:18.048445 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-k85cf"] Dec 01 22:37:18 crc kubenswrapper[4857]: I1201 22:37:18.068307 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-k85cf"] Dec 01 22:37:19 crc kubenswrapper[4857]: I1201 22:37:19.850109 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf9c958-5340-4c35-a0d5-fc5693c9f589" path="/var/lib/kubelet/pods/0bf9c958-5340-4c35-a0d5-fc5693c9f589/volumes" Dec 01 22:37:24 crc kubenswrapper[4857]: I1201 22:37:24.204685 4857 scope.go:117] "RemoveContainer" containerID="ce32b9add64c8ccbfc5d066896bab28739536bc89824ebc0cf77cae4ecc7e717" Dec 01 22:37:24 crc kubenswrapper[4857]: I1201 22:37:24.250666 4857 scope.go:117] "RemoveContainer" containerID="66ad3aa84195ee062595491b192c68262235c71f3a10bbfcb59c051fa27717e3" Dec 01 22:37:24 crc kubenswrapper[4857]: I1201 22:37:24.306746 4857 scope.go:117] "RemoveContainer" containerID="b24a0b8d6a1bb24674b263d6cc883d208a602e5c498f1252f0d2e318065873ce" Dec 01 22:37:57 crc kubenswrapper[4857]: I1201 22:37:57.806871 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:37:57 crc kubenswrapper[4857]: I1201 22:37:57.807723 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:38:24 crc kubenswrapper[4857]: I1201 22:38:24.489209 4857 scope.go:117] "RemoveContainer" containerID="5144a43c6f9c4eada7f058dcd0cbb90d3df519122f7d62fafebdd27443472174" Dec 01 22:38:24 crc kubenswrapper[4857]: I1201 22:38:24.545013 4857 scope.go:117] "RemoveContainer" containerID="3ecb47734c66430237ae9b314bd49ab2c834939b9edd79577cb35f70fab19150" Dec 01 22:38:24 crc kubenswrapper[4857]: I1201 22:38:24.577159 4857 scope.go:117] "RemoveContainer" containerID="e7928e5378f29db5337410c2925241367e1f79a6f6f038ad6b2966156971d56f" Dec 01 22:38:24 crc kubenswrapper[4857]: I1201 22:38:24.610145 4857 scope.go:117] "RemoveContainer" containerID="7f0cbeb943e426c39f61ced53e6f1266fc9a5661fa554d4a48415740ecd5d669" Dec 01 22:38:27 crc kubenswrapper[4857]: I1201 22:38:27.806226 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:38:27 crc kubenswrapper[4857]: I1201 22:38:27.806876 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:38:57 crc kubenswrapper[4857]: I1201 22:38:57.806033 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:38:57 crc kubenswrapper[4857]: I1201 22:38:57.806774 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:38:57 crc kubenswrapper[4857]: I1201 22:38:57.806842 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 22:38:57 crc kubenswrapper[4857]: I1201 22:38:57.807946 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 22:38:57 crc kubenswrapper[4857]: I1201 22:38:57.808084 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" gracePeriod=600 Dec 01 22:38:57 crc kubenswrapper[4857]: E1201 22:38:57.941774 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:38:58 crc kubenswrapper[4857]: I1201 22:38:58.250902 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" exitCode=0 Dec 01 22:38:58 crc kubenswrapper[4857]: I1201 22:38:58.250965 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104"} Dec 01 22:38:58 crc kubenswrapper[4857]: I1201 22:38:58.251006 4857 scope.go:117] "RemoveContainer" containerID="7018e5ef4fa6765cca44e184b68c5e2a16e99c3def48589b77a1494d6ddc3ee9" Dec 01 22:38:58 crc kubenswrapper[4857]: I1201 22:38:58.251812 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:38:58 crc kubenswrapper[4857]: E1201 22:38:58.252252 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:39:01 crc kubenswrapper[4857]: I1201 22:39:01.451140 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f765fc7d6-x797d_ff230e58-1711-4638-89d6-daf91fb3bfc1/manager/0.log" Dec 01 22:39:06 crc kubenswrapper[4857]: I1201 22:39:06.830997 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:39:06 crc kubenswrapper[4857]: I1201 22:39:06.832082 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="prometheus" containerID="cri-o://508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7" gracePeriod=600 Dec 01 22:39:06 crc kubenswrapper[4857]: I1201 22:39:06.832108 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="thanos-sidecar" containerID="cri-o://cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26" gracePeriod=600 Dec 01 22:39:06 crc kubenswrapper[4857]: I1201 22:39:06.832275 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="config-reloader" containerID="cri-o://eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d" gracePeriod=600 Dec 01 22:39:07 crc kubenswrapper[4857]: I1201 22:39:07.346799 4857 generic.go:334] "Generic (PLEG): container finished" podID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerID="cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26" exitCode=0 Dec 01 22:39:07 crc kubenswrapper[4857]: I1201 22:39:07.348523 4857 generic.go:334] "Generic (PLEG): container finished" podID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerID="508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7" exitCode=0 Dec 01 22:39:07 crc kubenswrapper[4857]: I1201 22:39:07.347330 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"84257e31-0d4a-4b6f-857f-fb396eb2faf7","Type":"ContainerDied","Data":"cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26"} Dec 01 22:39:07 crc kubenswrapper[4857]: I1201 22:39:07.348695 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"84257e31-0d4a-4b6f-857f-fb396eb2faf7","Type":"ContainerDied","Data":"508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7"} Dec 01 22:39:07 crc kubenswrapper[4857]: I1201 22:39:07.930165 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.048371 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config\") pod \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.048533 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/84257e31-0d4a-4b6f-857f-fb396eb2faf7-tls-assets\") pod \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.048620 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/84257e31-0d4a-4b6f-857f-fb396eb2faf7-config-out\") pod \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.048750 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59gtg\" (UniqueName: \"kubernetes.io/projected/84257e31-0d4a-4b6f-857f-fb396eb2faf7-kube-api-access-59gtg\") pod \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.048815 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/84257e31-0d4a-4b6f-857f-fb396eb2faf7-prometheus-metric-storage-rulefiles-0\") pod \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.048965 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/84257e31-0d4a-4b6f-857f-fb396eb2faf7-prometheus-metric-storage-db\") pod \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.049033 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-config\") pod \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.049147 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.049206 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-thanos-prometheus-http-client-file\") pod \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.049423 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-secret-combined-ca-bundle\") pod \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.049533 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\" (UID: \"84257e31-0d4a-4b6f-857f-fb396eb2faf7\") " Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.050467 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84257e31-0d4a-4b6f-857f-fb396eb2faf7-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "84257e31-0d4a-4b6f-857f-fb396eb2faf7" (UID: "84257e31-0d4a-4b6f-857f-fb396eb2faf7"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.051288 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84257e31-0d4a-4b6f-857f-fb396eb2faf7-prometheus-metric-storage-db" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "84257e31-0d4a-4b6f-857f-fb396eb2faf7" (UID: "84257e31-0d4a-4b6f-857f-fb396eb2faf7"). InnerVolumeSpecName "prometheus-metric-storage-db". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.056559 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84257e31-0d4a-4b6f-857f-fb396eb2faf7-kube-api-access-59gtg" (OuterVolumeSpecName: "kube-api-access-59gtg") pod "84257e31-0d4a-4b6f-857f-fb396eb2faf7" (UID: "84257e31-0d4a-4b6f-857f-fb396eb2faf7"). InnerVolumeSpecName "kube-api-access-59gtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.058896 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "84257e31-0d4a-4b6f-857f-fb396eb2faf7" (UID: "84257e31-0d4a-4b6f-857f-fb396eb2faf7"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.059000 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "84257e31-0d4a-4b6f-857f-fb396eb2faf7" (UID: "84257e31-0d4a-4b6f-857f-fb396eb2faf7"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.059131 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "84257e31-0d4a-4b6f-857f-fb396eb2faf7" (UID: "84257e31-0d4a-4b6f-857f-fb396eb2faf7"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.062852 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "84257e31-0d4a-4b6f-857f-fb396eb2faf7" (UID: "84257e31-0d4a-4b6f-857f-fb396eb2faf7"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.072027 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-config" (OuterVolumeSpecName: "config") pod "84257e31-0d4a-4b6f-857f-fb396eb2faf7" (UID: "84257e31-0d4a-4b6f-857f-fb396eb2faf7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.075744 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84257e31-0d4a-4b6f-857f-fb396eb2faf7-config-out" (OuterVolumeSpecName: "config-out") pod "84257e31-0d4a-4b6f-857f-fb396eb2faf7" (UID: "84257e31-0d4a-4b6f-857f-fb396eb2faf7"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.077788 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84257e31-0d4a-4b6f-857f-fb396eb2faf7-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "84257e31-0d4a-4b6f-857f-fb396eb2faf7" (UID: "84257e31-0d4a-4b6f-857f-fb396eb2faf7"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.141583 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config" (OuterVolumeSpecName: "web-config") pod "84257e31-0d4a-4b6f-857f-fb396eb2faf7" (UID: "84257e31-0d4a-4b6f-857f-fb396eb2faf7"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.152704 4857 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.152744 4857 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config\") on node \"crc\" DevicePath \"\"" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.152756 4857 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/84257e31-0d4a-4b6f-857f-fb396eb2faf7-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.152766 4857 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/84257e31-0d4a-4b6f-857f-fb396eb2faf7-config-out\") on node \"crc\" DevicePath \"\"" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.152777 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59gtg\" (UniqueName: \"kubernetes.io/projected/84257e31-0d4a-4b6f-857f-fb396eb2faf7-kube-api-access-59gtg\") on node \"crc\" DevicePath \"\"" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.152787 4857 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/84257e31-0d4a-4b6f-857f-fb396eb2faf7-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.152796 4857 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/84257e31-0d4a-4b6f-857f-fb396eb2faf7-prometheus-metric-storage-db\") on node \"crc\" DevicePath \"\"" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.152805 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-config\") on node \"crc\" DevicePath \"\"" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.152816 4857 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.152827 4857 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.152836 4857 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84257e31-0d4a-4b6f-857f-fb396eb2faf7-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.358877 4857 generic.go:334] "Generic (PLEG): container finished" podID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerID="eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d" exitCode=0 Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.358921 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"84257e31-0d4a-4b6f-857f-fb396eb2faf7","Type":"ContainerDied","Data":"eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d"} Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.358952 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"84257e31-0d4a-4b6f-857f-fb396eb2faf7","Type":"ContainerDied","Data":"34307c2f1fdecd7c2e1ed08ca099299af155bb57b719c9b0eb04f9f95ef4fecb"} Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.358971 4857 scope.go:117] "RemoveContainer" containerID="cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.358971 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.377686 4857 scope.go:117] "RemoveContainer" containerID="eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.413864 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.430208 4857 scope.go:117] "RemoveContainer" containerID="508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.442408 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.516990 4857 scope.go:117] "RemoveContainer" containerID="c3b46f9f80b298dd413728db6458bec4c1dc9e7e80d89d5095ff753ab0177ea8" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.569354 4857 scope.go:117] "RemoveContainer" containerID="cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26" Dec 01 22:39:08 crc kubenswrapper[4857]: E1201 22:39:08.569921 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26\": container with ID starting with cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26 not found: ID does not exist" containerID="cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.569973 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26"} err="failed to get container status \"cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26\": rpc error: code = NotFound desc = could not find container \"cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26\": container with ID starting with cb37fd3ec9b3c65a2e5a72a14eca384858f74c2ba4e59ba01872828272fbbf26 not found: ID does not exist" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.570000 4857 scope.go:117] "RemoveContainer" containerID="eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d" Dec 01 22:39:08 crc kubenswrapper[4857]: E1201 22:39:08.570374 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d\": container with ID starting with eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d not found: ID does not exist" containerID="eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.570426 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d"} err="failed to get container status \"eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d\": rpc error: code = NotFound desc = could not find container \"eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d\": container with ID starting with eefd0f47074c3eb6aa23fe6233703c550ce32b1b7c009bf2d1efe9eebdd3b97d not found: ID does not exist" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.570446 4857 scope.go:117] "RemoveContainer" containerID="508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7" Dec 01 22:39:08 crc kubenswrapper[4857]: E1201 22:39:08.570709 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7\": container with ID starting with 508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7 not found: ID does not exist" containerID="508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.570735 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7"} err="failed to get container status \"508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7\": rpc error: code = NotFound desc = could not find container \"508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7\": container with ID starting with 508a8e7617f0f2e308794d84d70eb77d1bd58c04c85bf74721e59793696acbf7 not found: ID does not exist" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.570750 4857 scope.go:117] "RemoveContainer" containerID="c3b46f9f80b298dd413728db6458bec4c1dc9e7e80d89d5095ff753ab0177ea8" Dec 01 22:39:08 crc kubenswrapper[4857]: E1201 22:39:08.571013 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3b46f9f80b298dd413728db6458bec4c1dc9e7e80d89d5095ff753ab0177ea8\": container with ID starting with c3b46f9f80b298dd413728db6458bec4c1dc9e7e80d89d5095ff753ab0177ea8 not found: ID does not exist" containerID="c3b46f9f80b298dd413728db6458bec4c1dc9e7e80d89d5095ff753ab0177ea8" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.571063 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3b46f9f80b298dd413728db6458bec4c1dc9e7e80d89d5095ff753ab0177ea8"} err="failed to get container status \"c3b46f9f80b298dd413728db6458bec4c1dc9e7e80d89d5095ff753ab0177ea8\": rpc error: code = NotFound desc = could not find container \"c3b46f9f80b298dd413728db6458bec4c1dc9e7e80d89d5095ff753ab0177ea8\": container with ID starting with c3b46f9f80b298dd413728db6458bec4c1dc9e7e80d89d5095ff753ab0177ea8 not found: ID does not exist" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.619825 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:39:08 crc kubenswrapper[4857]: E1201 22:39:08.620238 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="prometheus" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.620252 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="prometheus" Dec 01 22:39:08 crc kubenswrapper[4857]: E1201 22:39:08.620264 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="init-config-reloader" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.620272 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="init-config-reloader" Dec 01 22:39:08 crc kubenswrapper[4857]: E1201 22:39:08.620282 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="thanos-sidecar" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.620288 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="thanos-sidecar" Dec 01 22:39:08 crc kubenswrapper[4857]: E1201 22:39:08.620310 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="config-reloader" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.620316 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="config-reloader" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.620493 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="config-reloader" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.620510 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="prometheus" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.620525 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" containerName="thanos-sidecar" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.622206 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.627871 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-whwdt" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.628778 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.628962 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.629000 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.629503 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.632696 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.672928 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.701200 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.770519 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.770794 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.770904 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.771066 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-config\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.771142 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzwrg\" (UniqueName: \"kubernetes.io/projected/0203f043-2b36-4611-9b70-1fd54f97a1e3-kube-api-access-bzwrg\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.771246 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0203f043-2b36-4611-9b70-1fd54f97a1e3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.771288 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.771351 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0203f043-2b36-4611-9b70-1fd54f97a1e3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.771389 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/0203f043-2b36-4611-9b70-1fd54f97a1e3-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.771445 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.771470 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0203f043-2b36-4611-9b70-1fd54f97a1e3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.872750 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.872848 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0203f043-2b36-4611-9b70-1fd54f97a1e3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.872907 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/0203f043-2b36-4611-9b70-1fd54f97a1e3-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.873008 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.873233 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0203f043-2b36-4611-9b70-1fd54f97a1e3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.873422 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.873716 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/0203f043-2b36-4611-9b70-1fd54f97a1e3-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.873948 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.874137 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.874209 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-config\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.874246 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzwrg\" (UniqueName: \"kubernetes.io/projected/0203f043-2b36-4611-9b70-1fd54f97a1e3-kube-api-access-bzwrg\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.874311 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0203f043-2b36-4611-9b70-1fd54f97a1e3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.874714 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0203f043-2b36-4611-9b70-1fd54f97a1e3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.878562 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0203f043-2b36-4611-9b70-1fd54f97a1e3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.878609 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0203f043-2b36-4611-9b70-1fd54f97a1e3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.878938 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.878968 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.879149 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.879274 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.879764 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.880633 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0203f043-2b36-4611-9b70-1fd54f97a1e3-config\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.889562 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzwrg\" (UniqueName: \"kubernetes.io/projected/0203f043-2b36-4611-9b70-1fd54f97a1e3-kube-api-access-bzwrg\") pod \"prometheus-metric-storage-0\" (UID: \"0203f043-2b36-4611-9b70-1fd54f97a1e3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:08 crc kubenswrapper[4857]: I1201 22:39:08.938543 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:09 crc kubenswrapper[4857]: I1201 22:39:09.571164 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 22:39:09 crc kubenswrapper[4857]: I1201 22:39:09.850732 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84257e31-0d4a-4b6f-857f-fb396eb2faf7" path="/var/lib/kubelet/pods/84257e31-0d4a-4b6f-857f-fb396eb2faf7/volumes" Dec 01 22:39:10 crc kubenswrapper[4857]: I1201 22:39:10.399557 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0203f043-2b36-4611-9b70-1fd54f97a1e3","Type":"ContainerStarted","Data":"2bedbb10cc4fc1238d0e914e17f988f66e06d3450cfd65d4cc3d5eb356c277ce"} Dec 01 22:39:12 crc kubenswrapper[4857]: I1201 22:39:12.837240 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:39:12 crc kubenswrapper[4857]: E1201 22:39:12.838673 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:39:14 crc kubenswrapper[4857]: I1201 22:39:14.465085 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0203f043-2b36-4611-9b70-1fd54f97a1e3","Type":"ContainerStarted","Data":"01df9b512aa0cf2629d64315cd757a14d84149e0a58c3b9d4c26ae68c8a347cb"} Dec 01 22:39:23 crc kubenswrapper[4857]: I1201 22:39:23.835505 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:39:23 crc kubenswrapper[4857]: E1201 22:39:23.836617 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:39:25 crc kubenswrapper[4857]: I1201 22:39:25.609099 4857 generic.go:334] "Generic (PLEG): container finished" podID="0203f043-2b36-4611-9b70-1fd54f97a1e3" containerID="01df9b512aa0cf2629d64315cd757a14d84149e0a58c3b9d4c26ae68c8a347cb" exitCode=0 Dec 01 22:39:25 crc kubenswrapper[4857]: I1201 22:39:25.609312 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0203f043-2b36-4611-9b70-1fd54f97a1e3","Type":"ContainerDied","Data":"01df9b512aa0cf2629d64315cd757a14d84149e0a58c3b9d4c26ae68c8a347cb"} Dec 01 22:39:26 crc kubenswrapper[4857]: I1201 22:39:26.621689 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0203f043-2b36-4611-9b70-1fd54f97a1e3","Type":"ContainerStarted","Data":"c26feeb1d578eecee8b922e3555f33f269c0f323a2029af903bc85b87f908668"} Dec 01 22:39:32 crc kubenswrapper[4857]: I1201 22:39:32.706145 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0203f043-2b36-4611-9b70-1fd54f97a1e3","Type":"ContainerStarted","Data":"426d940b37db7f66b1854d3a0f683233f4bd4a6891cc27b300be36b9e3b89cba"} Dec 01 22:39:32 crc kubenswrapper[4857]: I1201 22:39:32.706779 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0203f043-2b36-4611-9b70-1fd54f97a1e3","Type":"ContainerStarted","Data":"73abdab4e941f8a8f1c8ce79d97fb3c1fd116d308a8bb0da7cf5adffd9031067"} Dec 01 22:39:32 crc kubenswrapper[4857]: I1201 22:39:32.746859 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=24.746840416 podStartE2EDuration="24.746840416s" podCreationTimestamp="2025-12-01 22:39:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 22:39:32.738796671 +0000 UTC m=+3891.228859008" watchObservedRunningTime="2025-12-01 22:39:32.746840416 +0000 UTC m=+3891.236902733" Dec 01 22:39:33 crc kubenswrapper[4857]: I1201 22:39:33.939087 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:35 crc kubenswrapper[4857]: I1201 22:39:35.835918 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:39:35 crc kubenswrapper[4857]: E1201 22:39:35.836651 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:39:37 crc kubenswrapper[4857]: I1201 22:39:37.907680 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kzffc"] Dec 01 22:39:37 crc kubenswrapper[4857]: I1201 22:39:37.912108 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:37 crc kubenswrapper[4857]: I1201 22:39:37.920141 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kzffc"] Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.049268 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-catalog-content\") pod \"redhat-operators-kzffc\" (UID: \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\") " pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.049456 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5m7t\" (UniqueName: \"kubernetes.io/projected/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-kube-api-access-g5m7t\") pod \"redhat-operators-kzffc\" (UID: \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\") " pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.049710 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-utilities\") pod \"redhat-operators-kzffc\" (UID: \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\") " pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.151701 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-utilities\") pod \"redhat-operators-kzffc\" (UID: \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\") " pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.151853 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-catalog-content\") pod \"redhat-operators-kzffc\" (UID: \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\") " pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.151911 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5m7t\" (UniqueName: \"kubernetes.io/projected/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-kube-api-access-g5m7t\") pod \"redhat-operators-kzffc\" (UID: \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\") " pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.152859 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-utilities\") pod \"redhat-operators-kzffc\" (UID: \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\") " pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.153227 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-catalog-content\") pod \"redhat-operators-kzffc\" (UID: \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\") " pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.173650 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5m7t\" (UniqueName: \"kubernetes.io/projected/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-kube-api-access-g5m7t\") pod \"redhat-operators-kzffc\" (UID: \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\") " pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.254677 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.799150 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kzffc"] Dec 01 22:39:38 crc kubenswrapper[4857]: W1201 22:39:38.806297 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b232d0c_3b18_43df_9c6d_998d7bab1b9a.slice/crio-1c5d0403c95c5ad535a5ede8e9318336ca4e8d3fc2ad3f43ebbcf4783bb787ed WatchSource:0}: Error finding container 1c5d0403c95c5ad535a5ede8e9318336ca4e8d3fc2ad3f43ebbcf4783bb787ed: Status 404 returned error can't find the container with id 1c5d0403c95c5ad535a5ede8e9318336ca4e8d3fc2ad3f43ebbcf4783bb787ed Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.939126 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:38 crc kubenswrapper[4857]: I1201 22:39:38.945801 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:39 crc kubenswrapper[4857]: I1201 22:39:39.798198 4857 generic.go:334] "Generic (PLEG): container finished" podID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" containerID="299877b27446c3aa3c94bec711f2d2fb5625bdf3a53d3a4ccf84489810aa348c" exitCode=0 Dec 01 22:39:39 crc kubenswrapper[4857]: I1201 22:39:39.798828 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzffc" event={"ID":"8b232d0c-3b18-43df-9c6d-998d7bab1b9a","Type":"ContainerDied","Data":"299877b27446c3aa3c94bec711f2d2fb5625bdf3a53d3a4ccf84489810aa348c"} Dec 01 22:39:39 crc kubenswrapper[4857]: I1201 22:39:39.799092 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzffc" event={"ID":"8b232d0c-3b18-43df-9c6d-998d7bab1b9a","Type":"ContainerStarted","Data":"1c5d0403c95c5ad535a5ede8e9318336ca4e8d3fc2ad3f43ebbcf4783bb787ed"} Dec 01 22:39:39 crc kubenswrapper[4857]: I1201 22:39:39.806894 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 01 22:39:41 crc kubenswrapper[4857]: I1201 22:39:41.883751 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzffc" event={"ID":"8b232d0c-3b18-43df-9c6d-998d7bab1b9a","Type":"ContainerStarted","Data":"e0eae5810b6c210dcefa0882bf3c2c267dfee6a74ae09e12c6c93ab7fc4f2db2"} Dec 01 22:39:44 crc kubenswrapper[4857]: I1201 22:39:44.877617 4857 generic.go:334] "Generic (PLEG): container finished" podID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" containerID="e0eae5810b6c210dcefa0882bf3c2c267dfee6a74ae09e12c6c93ab7fc4f2db2" exitCode=0 Dec 01 22:39:44 crc kubenswrapper[4857]: I1201 22:39:44.877713 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzffc" event={"ID":"8b232d0c-3b18-43df-9c6d-998d7bab1b9a","Type":"ContainerDied","Data":"e0eae5810b6c210dcefa0882bf3c2c267dfee6a74ae09e12c6c93ab7fc4f2db2"} Dec 01 22:39:46 crc kubenswrapper[4857]: I1201 22:39:46.926503 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzffc" event={"ID":"8b232d0c-3b18-43df-9c6d-998d7bab1b9a","Type":"ContainerStarted","Data":"f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21"} Dec 01 22:39:46 crc kubenswrapper[4857]: I1201 22:39:46.962960 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kzffc" podStartSLOduration=4.209370599 podStartE2EDuration="9.96293196s" podCreationTimestamp="2025-12-01 22:39:37 +0000 UTC" firstStartedPulling="2025-12-01 22:39:39.804099649 +0000 UTC m=+3898.294162006" lastFinishedPulling="2025-12-01 22:39:45.55766105 +0000 UTC m=+3904.047723367" observedRunningTime="2025-12-01 22:39:46.952724292 +0000 UTC m=+3905.442786609" watchObservedRunningTime="2025-12-01 22:39:46.96293196 +0000 UTC m=+3905.452994287" Dec 01 22:39:48 crc kubenswrapper[4857]: I1201 22:39:48.255469 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:48 crc kubenswrapper[4857]: I1201 22:39:48.255804 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:49 crc kubenswrapper[4857]: I1201 22:39:49.348741 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kzffc" podUID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" containerName="registry-server" probeResult="failure" output=< Dec 01 22:39:49 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Dec 01 22:39:49 crc kubenswrapper[4857]: > Dec 01 22:39:50 crc kubenswrapper[4857]: I1201 22:39:50.835071 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:39:50 crc kubenswrapper[4857]: E1201 22:39:50.835595 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:39:58 crc kubenswrapper[4857]: I1201 22:39:58.370695 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:58 crc kubenswrapper[4857]: I1201 22:39:58.445103 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:39:58 crc kubenswrapper[4857]: I1201 22:39:58.638425 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kzffc"] Dec 01 22:40:00 crc kubenswrapper[4857]: I1201 22:40:00.074430 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kzffc" podUID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" containerName="registry-server" containerID="cri-o://f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21" gracePeriod=2 Dec 01 22:40:00 crc kubenswrapper[4857]: I1201 22:40:00.717669 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:40:00 crc kubenswrapper[4857]: I1201 22:40:00.809861 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-catalog-content\") pod \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\" (UID: \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\") " Dec 01 22:40:00 crc kubenswrapper[4857]: I1201 22:40:00.810004 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-utilities\") pod \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\" (UID: \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\") " Dec 01 22:40:00 crc kubenswrapper[4857]: I1201 22:40:00.810135 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5m7t\" (UniqueName: \"kubernetes.io/projected/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-kube-api-access-g5m7t\") pod \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\" (UID: \"8b232d0c-3b18-43df-9c6d-998d7bab1b9a\") " Dec 01 22:40:00 crc kubenswrapper[4857]: I1201 22:40:00.810857 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-utilities" (OuterVolumeSpecName: "utilities") pod "8b232d0c-3b18-43df-9c6d-998d7bab1b9a" (UID: "8b232d0c-3b18-43df-9c6d-998d7bab1b9a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:40:00 crc kubenswrapper[4857]: I1201 22:40:00.913687 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:40:00 crc kubenswrapper[4857]: I1201 22:40:00.941999 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b232d0c-3b18-43df-9c6d-998d7bab1b9a" (UID: "8b232d0c-3b18-43df-9c6d-998d7bab1b9a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.016664 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.089393 4857 generic.go:334] "Generic (PLEG): container finished" podID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" containerID="f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21" exitCode=0 Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.089454 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzffc" event={"ID":"8b232d0c-3b18-43df-9c6d-998d7bab1b9a","Type":"ContainerDied","Data":"f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21"} Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.089486 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kzffc" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.089502 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzffc" event={"ID":"8b232d0c-3b18-43df-9c6d-998d7bab1b9a","Type":"ContainerDied","Data":"1c5d0403c95c5ad535a5ede8e9318336ca4e8d3fc2ad3f43ebbcf4783bb787ed"} Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.089534 4857 scope.go:117] "RemoveContainer" containerID="f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.136390 4857 scope.go:117] "RemoveContainer" containerID="e0eae5810b6c210dcefa0882bf3c2c267dfee6a74ae09e12c6c93ab7fc4f2db2" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.463935 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-kube-api-access-g5m7t" (OuterVolumeSpecName: "kube-api-access-g5m7t") pod "8b232d0c-3b18-43df-9c6d-998d7bab1b9a" (UID: "8b232d0c-3b18-43df-9c6d-998d7bab1b9a"). InnerVolumeSpecName "kube-api-access-g5m7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.502863 4857 scope.go:117] "RemoveContainer" containerID="299877b27446c3aa3c94bec711f2d2fb5625bdf3a53d3a4ccf84489810aa348c" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.558719 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5m7t\" (UniqueName: \"kubernetes.io/projected/8b232d0c-3b18-43df-9c6d-998d7bab1b9a-kube-api-access-g5m7t\") on node \"crc\" DevicePath \"\"" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.585236 4857 scope.go:117] "RemoveContainer" containerID="f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21" Dec 01 22:40:01 crc kubenswrapper[4857]: E1201 22:40:01.586392 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21\": container with ID starting with f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21 not found: ID does not exist" containerID="f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.586422 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21"} err="failed to get container status \"f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21\": rpc error: code = NotFound desc = could not find container \"f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21\": container with ID starting with f2dc451dfaa9e42f4322fc37d87aed7640d08a7818c62097aeea0b2ebf093d21 not found: ID does not exist" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.586443 4857 scope.go:117] "RemoveContainer" containerID="e0eae5810b6c210dcefa0882bf3c2c267dfee6a74ae09e12c6c93ab7fc4f2db2" Dec 01 22:40:01 crc kubenswrapper[4857]: E1201 22:40:01.587023 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0eae5810b6c210dcefa0882bf3c2c267dfee6a74ae09e12c6c93ab7fc4f2db2\": container with ID starting with e0eae5810b6c210dcefa0882bf3c2c267dfee6a74ae09e12c6c93ab7fc4f2db2 not found: ID does not exist" containerID="e0eae5810b6c210dcefa0882bf3c2c267dfee6a74ae09e12c6c93ab7fc4f2db2" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.587055 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0eae5810b6c210dcefa0882bf3c2c267dfee6a74ae09e12c6c93ab7fc4f2db2"} err="failed to get container status \"e0eae5810b6c210dcefa0882bf3c2c267dfee6a74ae09e12c6c93ab7fc4f2db2\": rpc error: code = NotFound desc = could not find container \"e0eae5810b6c210dcefa0882bf3c2c267dfee6a74ae09e12c6c93ab7fc4f2db2\": container with ID starting with e0eae5810b6c210dcefa0882bf3c2c267dfee6a74ae09e12c6c93ab7fc4f2db2 not found: ID does not exist" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.587072 4857 scope.go:117] "RemoveContainer" containerID="299877b27446c3aa3c94bec711f2d2fb5625bdf3a53d3a4ccf84489810aa348c" Dec 01 22:40:01 crc kubenswrapper[4857]: E1201 22:40:01.587439 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"299877b27446c3aa3c94bec711f2d2fb5625bdf3a53d3a4ccf84489810aa348c\": container with ID starting with 299877b27446c3aa3c94bec711f2d2fb5625bdf3a53d3a4ccf84489810aa348c not found: ID does not exist" containerID="299877b27446c3aa3c94bec711f2d2fb5625bdf3a53d3a4ccf84489810aa348c" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.587463 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"299877b27446c3aa3c94bec711f2d2fb5625bdf3a53d3a4ccf84489810aa348c"} err="failed to get container status \"299877b27446c3aa3c94bec711f2d2fb5625bdf3a53d3a4ccf84489810aa348c\": rpc error: code = NotFound desc = could not find container \"299877b27446c3aa3c94bec711f2d2fb5625bdf3a53d3a4ccf84489810aa348c\": container with ID starting with 299877b27446c3aa3c94bec711f2d2fb5625bdf3a53d3a4ccf84489810aa348c not found: ID does not exist" Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.742610 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kzffc"] Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.754026 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kzffc"] Dec 01 22:40:01 crc kubenswrapper[4857]: I1201 22:40:01.845956 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" path="/var/lib/kubelet/pods/8b232d0c-3b18-43df-9c6d-998d7bab1b9a/volumes" Dec 01 22:40:03 crc kubenswrapper[4857]: I1201 22:40:03.835325 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:40:03 crc kubenswrapper[4857]: E1201 22:40:03.836184 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:40:09 crc kubenswrapper[4857]: I1201 22:40:09.814168 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kgqh9"] Dec 01 22:40:09 crc kubenswrapper[4857]: E1201 22:40:09.816427 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" containerName="registry-server" Dec 01 22:40:09 crc kubenswrapper[4857]: I1201 22:40:09.816778 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" containerName="registry-server" Dec 01 22:40:09 crc kubenswrapper[4857]: E1201 22:40:09.816863 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" containerName="extract-utilities" Dec 01 22:40:09 crc kubenswrapper[4857]: I1201 22:40:09.816889 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" containerName="extract-utilities" Dec 01 22:40:09 crc kubenswrapper[4857]: E1201 22:40:09.816942 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" containerName="extract-content" Dec 01 22:40:09 crc kubenswrapper[4857]: I1201 22:40:09.816977 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" containerName="extract-content" Dec 01 22:40:09 crc kubenswrapper[4857]: I1201 22:40:09.817860 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b232d0c-3b18-43df-9c6d-998d7bab1b9a" containerName="registry-server" Dec 01 22:40:09 crc kubenswrapper[4857]: I1201 22:40:09.982666 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:09 crc kubenswrapper[4857]: I1201 22:40:09.983188 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b689b746-de83-4849-9a5f-3c368704dced-catalog-content\") pod \"certified-operators-kgqh9\" (UID: \"b689b746-de83-4849-9a5f-3c368704dced\") " pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:09 crc kubenswrapper[4857]: I1201 22:40:09.983363 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b689b746-de83-4849-9a5f-3c368704dced-utilities\") pod \"certified-operators-kgqh9\" (UID: \"b689b746-de83-4849-9a5f-3c368704dced\") " pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:09 crc kubenswrapper[4857]: I1201 22:40:09.983403 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2ph7\" (UniqueName: \"kubernetes.io/projected/b689b746-de83-4849-9a5f-3c368704dced-kube-api-access-r2ph7\") pod \"certified-operators-kgqh9\" (UID: \"b689b746-de83-4849-9a5f-3c368704dced\") " pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.014672 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kgqh9"] Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.052168 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f852l"] Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.054218 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.066096 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f852l"] Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.088626 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3448686-7358-45a7-bd12-26d455bea6ff-utilities\") pod \"community-operators-f852l\" (UID: \"b3448686-7358-45a7-bd12-26d455bea6ff\") " pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.088722 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3448686-7358-45a7-bd12-26d455bea6ff-catalog-content\") pod \"community-operators-f852l\" (UID: \"b3448686-7358-45a7-bd12-26d455bea6ff\") " pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.088801 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b689b746-de83-4849-9a5f-3c368704dced-utilities\") pod \"certified-operators-kgqh9\" (UID: \"b689b746-de83-4849-9a5f-3c368704dced\") " pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.088827 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2ph7\" (UniqueName: \"kubernetes.io/projected/b689b746-de83-4849-9a5f-3c368704dced-kube-api-access-r2ph7\") pod \"certified-operators-kgqh9\" (UID: \"b689b746-de83-4849-9a5f-3c368704dced\") " pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.088973 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk22p\" (UniqueName: \"kubernetes.io/projected/b3448686-7358-45a7-bd12-26d455bea6ff-kube-api-access-rk22p\") pod \"community-operators-f852l\" (UID: \"b3448686-7358-45a7-bd12-26d455bea6ff\") " pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.089004 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b689b746-de83-4849-9a5f-3c368704dced-catalog-content\") pod \"certified-operators-kgqh9\" (UID: \"b689b746-de83-4849-9a5f-3c368704dced\") " pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.089609 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b689b746-de83-4849-9a5f-3c368704dced-catalog-content\") pod \"certified-operators-kgqh9\" (UID: \"b689b746-de83-4849-9a5f-3c368704dced\") " pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.092416 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b689b746-de83-4849-9a5f-3c368704dced-utilities\") pod \"certified-operators-kgqh9\" (UID: \"b689b746-de83-4849-9a5f-3c368704dced\") " pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.110805 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2ph7\" (UniqueName: \"kubernetes.io/projected/b689b746-de83-4849-9a5f-3c368704dced-kube-api-access-r2ph7\") pod \"certified-operators-kgqh9\" (UID: \"b689b746-de83-4849-9a5f-3c368704dced\") " pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.190707 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3448686-7358-45a7-bd12-26d455bea6ff-utilities\") pod \"community-operators-f852l\" (UID: \"b3448686-7358-45a7-bd12-26d455bea6ff\") " pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.190799 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3448686-7358-45a7-bd12-26d455bea6ff-catalog-content\") pod \"community-operators-f852l\" (UID: \"b3448686-7358-45a7-bd12-26d455bea6ff\") " pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.190960 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk22p\" (UniqueName: \"kubernetes.io/projected/b3448686-7358-45a7-bd12-26d455bea6ff-kube-api-access-rk22p\") pod \"community-operators-f852l\" (UID: \"b3448686-7358-45a7-bd12-26d455bea6ff\") " pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.191401 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3448686-7358-45a7-bd12-26d455bea6ff-utilities\") pod \"community-operators-f852l\" (UID: \"b3448686-7358-45a7-bd12-26d455bea6ff\") " pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.191428 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3448686-7358-45a7-bd12-26d455bea6ff-catalog-content\") pod \"community-operators-f852l\" (UID: \"b3448686-7358-45a7-bd12-26d455bea6ff\") " pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.215877 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk22p\" (UniqueName: \"kubernetes.io/projected/b3448686-7358-45a7-bd12-26d455bea6ff-kube-api-access-rk22p\") pod \"community-operators-f852l\" (UID: \"b3448686-7358-45a7-bd12-26d455bea6ff\") " pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.333392 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:10 crc kubenswrapper[4857]: I1201 22:40:10.371573 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:11 crc kubenswrapper[4857]: I1201 22:40:11.001065 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f852l"] Dec 01 22:40:11 crc kubenswrapper[4857]: I1201 22:40:11.009681 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kgqh9"] Dec 01 22:40:11 crc kubenswrapper[4857]: I1201 22:40:11.201352 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgqh9" event={"ID":"b689b746-de83-4849-9a5f-3c368704dced","Type":"ContainerStarted","Data":"38ad504178b3d176ee022bd2b59008556615b481264e2ce055005ea6eabe7af7"} Dec 01 22:40:11 crc kubenswrapper[4857]: I1201 22:40:11.202403 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f852l" event={"ID":"b3448686-7358-45a7-bd12-26d455bea6ff","Type":"ContainerStarted","Data":"6b9379b064f232bf25cfc28ec8813019513d89461f195e0f7a11cf4c58356985"} Dec 01 22:40:12 crc kubenswrapper[4857]: I1201 22:40:12.219674 4857 generic.go:334] "Generic (PLEG): container finished" podID="b689b746-de83-4849-9a5f-3c368704dced" containerID="3e74e87797fead10166071b90b0ec276340a4083a1b5d167a4786122e94f347f" exitCode=0 Dec 01 22:40:12 crc kubenswrapper[4857]: I1201 22:40:12.220395 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgqh9" event={"ID":"b689b746-de83-4849-9a5f-3c368704dced","Type":"ContainerDied","Data":"3e74e87797fead10166071b90b0ec276340a4083a1b5d167a4786122e94f347f"} Dec 01 22:40:12 crc kubenswrapper[4857]: I1201 22:40:12.226344 4857 generic.go:334] "Generic (PLEG): container finished" podID="b3448686-7358-45a7-bd12-26d455bea6ff" containerID="d863031120cce0a0ac5b5f1e6227c02245a2aabd7a8bed6a73444d064f3bc3b3" exitCode=0 Dec 01 22:40:12 crc kubenswrapper[4857]: I1201 22:40:12.226390 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f852l" event={"ID":"b3448686-7358-45a7-bd12-26d455bea6ff","Type":"ContainerDied","Data":"d863031120cce0a0ac5b5f1e6227c02245a2aabd7a8bed6a73444d064f3bc3b3"} Dec 01 22:40:14 crc kubenswrapper[4857]: I1201 22:40:14.337442 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgqh9" event={"ID":"b689b746-de83-4849-9a5f-3c368704dced","Type":"ContainerStarted","Data":"e99064a1c176b3ec2cf0759fc95c4ec8fcee1863aefb874f5395c26edac58ae9"} Dec 01 22:40:14 crc kubenswrapper[4857]: I1201 22:40:14.344267 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f852l" event={"ID":"b3448686-7358-45a7-bd12-26d455bea6ff","Type":"ContainerStarted","Data":"53f79f88d4a9d46b1ba7cb609652a21f453e916a62998a5eb34f89ceb25e4406"} Dec 01 22:40:16 crc kubenswrapper[4857]: I1201 22:40:16.368819 4857 generic.go:334] "Generic (PLEG): container finished" podID="b3448686-7358-45a7-bd12-26d455bea6ff" containerID="53f79f88d4a9d46b1ba7cb609652a21f453e916a62998a5eb34f89ceb25e4406" exitCode=0 Dec 01 22:40:16 crc kubenswrapper[4857]: I1201 22:40:16.369380 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f852l" event={"ID":"b3448686-7358-45a7-bd12-26d455bea6ff","Type":"ContainerDied","Data":"53f79f88d4a9d46b1ba7cb609652a21f453e916a62998a5eb34f89ceb25e4406"} Dec 01 22:40:16 crc kubenswrapper[4857]: I1201 22:40:16.835980 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:40:16 crc kubenswrapper[4857]: E1201 22:40:16.836350 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:40:20 crc kubenswrapper[4857]: I1201 22:40:20.423198 4857 generic.go:334] "Generic (PLEG): container finished" podID="b689b746-de83-4849-9a5f-3c368704dced" containerID="e99064a1c176b3ec2cf0759fc95c4ec8fcee1863aefb874f5395c26edac58ae9" exitCode=0 Dec 01 22:40:20 crc kubenswrapper[4857]: I1201 22:40:20.423239 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgqh9" event={"ID":"b689b746-de83-4849-9a5f-3c368704dced","Type":"ContainerDied","Data":"e99064a1c176b3ec2cf0759fc95c4ec8fcee1863aefb874f5395c26edac58ae9"} Dec 01 22:40:20 crc kubenswrapper[4857]: I1201 22:40:20.426205 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f852l" event={"ID":"b3448686-7358-45a7-bd12-26d455bea6ff","Type":"ContainerStarted","Data":"217ac4c6e440fb69f5df2960e53538bb9a504d39846d101b7f0af04d9c213453"} Dec 01 22:40:20 crc kubenswrapper[4857]: I1201 22:40:20.476296 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f852l" podStartSLOduration=6.866296437 podStartE2EDuration="11.476257076s" podCreationTimestamp="2025-12-01 22:40:09 +0000 UTC" firstStartedPulling="2025-12-01 22:40:12.232392977 +0000 UTC m=+3930.722455324" lastFinishedPulling="2025-12-01 22:40:16.842353636 +0000 UTC m=+3935.332415963" observedRunningTime="2025-12-01 22:40:20.468545859 +0000 UTC m=+3938.958608186" watchObservedRunningTime="2025-12-01 22:40:20.476257076 +0000 UTC m=+3938.966319393" Dec 01 22:40:21 crc kubenswrapper[4857]: I1201 22:40:21.440016 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgqh9" event={"ID":"b689b746-de83-4849-9a5f-3c368704dced","Type":"ContainerStarted","Data":"35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11"} Dec 01 22:40:21 crc kubenswrapper[4857]: I1201 22:40:21.473253 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kgqh9" podStartSLOduration=3.603734291 podStartE2EDuration="12.473225693s" podCreationTimestamp="2025-12-01 22:40:09 +0000 UTC" firstStartedPulling="2025-12-01 22:40:12.225758816 +0000 UTC m=+3930.715821173" lastFinishedPulling="2025-12-01 22:40:21.095250258 +0000 UTC m=+3939.585312575" observedRunningTime="2025-12-01 22:40:21.457204914 +0000 UTC m=+3939.947267231" watchObservedRunningTime="2025-12-01 22:40:21.473225693 +0000 UTC m=+3939.963288040" Dec 01 22:40:30 crc kubenswrapper[4857]: I1201 22:40:30.334507 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:30 crc kubenswrapper[4857]: I1201 22:40:30.335273 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:30 crc kubenswrapper[4857]: I1201 22:40:30.371874 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:30 crc kubenswrapper[4857]: I1201 22:40:30.371929 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:30 crc kubenswrapper[4857]: I1201 22:40:30.421674 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:30 crc kubenswrapper[4857]: I1201 22:40:30.462280 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:30 crc kubenswrapper[4857]: I1201 22:40:30.708299 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:30 crc kubenswrapper[4857]: I1201 22:40:30.710648 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:31 crc kubenswrapper[4857]: I1201 22:40:31.847116 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:40:31 crc kubenswrapper[4857]: E1201 22:40:31.847426 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:40:32 crc kubenswrapper[4857]: I1201 22:40:32.482201 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kgqh9"] Dec 01 22:40:32 crc kubenswrapper[4857]: I1201 22:40:32.663992 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kgqh9" podUID="b689b746-de83-4849-9a5f-3c368704dced" containerName="registry-server" containerID="cri-o://35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11" gracePeriod=2 Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.095942 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f852l"] Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.096513 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f852l" podUID="b3448686-7358-45a7-bd12-26d455bea6ff" containerName="registry-server" containerID="cri-o://217ac4c6e440fb69f5df2960e53538bb9a504d39846d101b7f0af04d9c213453" gracePeriod=2 Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.369440 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.489213 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2ph7\" (UniqueName: \"kubernetes.io/projected/b689b746-de83-4849-9a5f-3c368704dced-kube-api-access-r2ph7\") pod \"b689b746-de83-4849-9a5f-3c368704dced\" (UID: \"b689b746-de83-4849-9a5f-3c368704dced\") " Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.489615 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b689b746-de83-4849-9a5f-3c368704dced-utilities\") pod \"b689b746-de83-4849-9a5f-3c368704dced\" (UID: \"b689b746-de83-4849-9a5f-3c368704dced\") " Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.489763 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b689b746-de83-4849-9a5f-3c368704dced-catalog-content\") pod \"b689b746-de83-4849-9a5f-3c368704dced\" (UID: \"b689b746-de83-4849-9a5f-3c368704dced\") " Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.492698 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b689b746-de83-4849-9a5f-3c368704dced-utilities" (OuterVolumeSpecName: "utilities") pod "b689b746-de83-4849-9a5f-3c368704dced" (UID: "b689b746-de83-4849-9a5f-3c368704dced"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.544686 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b689b746-de83-4849-9a5f-3c368704dced-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b689b746-de83-4849-9a5f-3c368704dced" (UID: "b689b746-de83-4849-9a5f-3c368704dced"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.592221 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b689b746-de83-4849-9a5f-3c368704dced-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.592249 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b689b746-de83-4849-9a5f-3c368704dced-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.684670 4857 generic.go:334] "Generic (PLEG): container finished" podID="b689b746-de83-4849-9a5f-3c368704dced" containerID="35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11" exitCode=0 Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.684771 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgqh9" event={"ID":"b689b746-de83-4849-9a5f-3c368704dced","Type":"ContainerDied","Data":"35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11"} Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.684820 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgqh9" event={"ID":"b689b746-de83-4849-9a5f-3c368704dced","Type":"ContainerDied","Data":"38ad504178b3d176ee022bd2b59008556615b481264e2ce055005ea6eabe7af7"} Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.684852 4857 scope.go:117] "RemoveContainer" containerID="35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11" Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.685126 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kgqh9" Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.693564 4857 generic.go:334] "Generic (PLEG): container finished" podID="b3448686-7358-45a7-bd12-26d455bea6ff" containerID="217ac4c6e440fb69f5df2960e53538bb9a504d39846d101b7f0af04d9c213453" exitCode=0 Dec 01 22:40:33 crc kubenswrapper[4857]: I1201 22:40:33.693630 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f852l" event={"ID":"b3448686-7358-45a7-bd12-26d455bea6ff","Type":"ContainerDied","Data":"217ac4c6e440fb69f5df2960e53538bb9a504d39846d101b7f0af04d9c213453"} Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.257628 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b689b746-de83-4849-9a5f-3c368704dced-kube-api-access-r2ph7" (OuterVolumeSpecName: "kube-api-access-r2ph7") pod "b689b746-de83-4849-9a5f-3c368704dced" (UID: "b689b746-de83-4849-9a5f-3c368704dced"). InnerVolumeSpecName "kube-api-access-r2ph7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.283845 4857 scope.go:117] "RemoveContainer" containerID="e99064a1c176b3ec2cf0759fc95c4ec8fcee1863aefb874f5395c26edac58ae9" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.306887 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2ph7\" (UniqueName: \"kubernetes.io/projected/b689b746-de83-4849-9a5f-3c368704dced-kube-api-access-r2ph7\") on node \"crc\" DevicePath \"\"" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.609878 4857 scope.go:117] "RemoveContainer" containerID="3e74e87797fead10166071b90b0ec276340a4083a1b5d167a4786122e94f347f" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.626948 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.655634 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kgqh9"] Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.663407 4857 scope.go:117] "RemoveContainer" containerID="35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11" Dec 01 22:40:34 crc kubenswrapper[4857]: E1201 22:40:34.664471 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11\": container with ID starting with 35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11 not found: ID does not exist" containerID="35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.664531 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11"} err="failed to get container status \"35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11\": rpc error: code = NotFound desc = could not find container \"35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11\": container with ID starting with 35ba7a92506bb7ce574e648f3d27a8809b0c1e2b97f557421e735038e10f4f11 not found: ID does not exist" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.664566 4857 scope.go:117] "RemoveContainer" containerID="e99064a1c176b3ec2cf0759fc95c4ec8fcee1863aefb874f5395c26edac58ae9" Dec 01 22:40:34 crc kubenswrapper[4857]: E1201 22:40:34.666801 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e99064a1c176b3ec2cf0759fc95c4ec8fcee1863aefb874f5395c26edac58ae9\": container with ID starting with e99064a1c176b3ec2cf0759fc95c4ec8fcee1863aefb874f5395c26edac58ae9 not found: ID does not exist" containerID="e99064a1c176b3ec2cf0759fc95c4ec8fcee1863aefb874f5395c26edac58ae9" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.667099 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e99064a1c176b3ec2cf0759fc95c4ec8fcee1863aefb874f5395c26edac58ae9"} err="failed to get container status \"e99064a1c176b3ec2cf0759fc95c4ec8fcee1863aefb874f5395c26edac58ae9\": rpc error: code = NotFound desc = could not find container \"e99064a1c176b3ec2cf0759fc95c4ec8fcee1863aefb874f5395c26edac58ae9\": container with ID starting with e99064a1c176b3ec2cf0759fc95c4ec8fcee1863aefb874f5395c26edac58ae9 not found: ID does not exist" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.667151 4857 scope.go:117] "RemoveContainer" containerID="3e74e87797fead10166071b90b0ec276340a4083a1b5d167a4786122e94f347f" Dec 01 22:40:34 crc kubenswrapper[4857]: E1201 22:40:34.669182 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e74e87797fead10166071b90b0ec276340a4083a1b5d167a4786122e94f347f\": container with ID starting with 3e74e87797fead10166071b90b0ec276340a4083a1b5d167a4786122e94f347f not found: ID does not exist" containerID="3e74e87797fead10166071b90b0ec276340a4083a1b5d167a4786122e94f347f" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.669249 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e74e87797fead10166071b90b0ec276340a4083a1b5d167a4786122e94f347f"} err="failed to get container status \"3e74e87797fead10166071b90b0ec276340a4083a1b5d167a4786122e94f347f\": rpc error: code = NotFound desc = could not find container \"3e74e87797fead10166071b90b0ec276340a4083a1b5d167a4786122e94f347f\": container with ID starting with 3e74e87797fead10166071b90b0ec276340a4083a1b5d167a4786122e94f347f not found: ID does not exist" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.706452 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f852l" event={"ID":"b3448686-7358-45a7-bd12-26d455bea6ff","Type":"ContainerDied","Data":"6b9379b064f232bf25cfc28ec8813019513d89461f195e0f7a11cf4c58356985"} Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.706733 4857 scope.go:117] "RemoveContainer" containerID="217ac4c6e440fb69f5df2960e53538bb9a504d39846d101b7f0af04d9c213453" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.706937 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f852l" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.715507 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kgqh9"] Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.716648 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3448686-7358-45a7-bd12-26d455bea6ff-catalog-content\") pod \"b3448686-7358-45a7-bd12-26d455bea6ff\" (UID: \"b3448686-7358-45a7-bd12-26d455bea6ff\") " Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.716805 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rk22p\" (UniqueName: \"kubernetes.io/projected/b3448686-7358-45a7-bd12-26d455bea6ff-kube-api-access-rk22p\") pod \"b3448686-7358-45a7-bd12-26d455bea6ff\" (UID: \"b3448686-7358-45a7-bd12-26d455bea6ff\") " Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.716957 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3448686-7358-45a7-bd12-26d455bea6ff-utilities\") pod \"b3448686-7358-45a7-bd12-26d455bea6ff\" (UID: \"b3448686-7358-45a7-bd12-26d455bea6ff\") " Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.720977 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3448686-7358-45a7-bd12-26d455bea6ff-utilities" (OuterVolumeSpecName: "utilities") pod "b3448686-7358-45a7-bd12-26d455bea6ff" (UID: "b3448686-7358-45a7-bd12-26d455bea6ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.726344 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3448686-7358-45a7-bd12-26d455bea6ff-kube-api-access-rk22p" (OuterVolumeSpecName: "kube-api-access-rk22p") pod "b3448686-7358-45a7-bd12-26d455bea6ff" (UID: "b3448686-7358-45a7-bd12-26d455bea6ff"). InnerVolumeSpecName "kube-api-access-rk22p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.742716 4857 scope.go:117] "RemoveContainer" containerID="53f79f88d4a9d46b1ba7cb609652a21f453e916a62998a5eb34f89ceb25e4406" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.782867 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3448686-7358-45a7-bd12-26d455bea6ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3448686-7358-45a7-bd12-26d455bea6ff" (UID: "b3448686-7358-45a7-bd12-26d455bea6ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.788908 4857 scope.go:117] "RemoveContainer" containerID="d863031120cce0a0ac5b5f1e6227c02245a2aabd7a8bed6a73444d064f3bc3b3" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.819028 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3448686-7358-45a7-bd12-26d455bea6ff-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.819063 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rk22p\" (UniqueName: \"kubernetes.io/projected/b3448686-7358-45a7-bd12-26d455bea6ff-kube-api-access-rk22p\") on node \"crc\" DevicePath \"\"" Dec 01 22:40:34 crc kubenswrapper[4857]: I1201 22:40:34.819076 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3448686-7358-45a7-bd12-26d455bea6ff-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:40:35 crc kubenswrapper[4857]: I1201 22:40:35.073814 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f852l"] Dec 01 22:40:35 crc kubenswrapper[4857]: I1201 22:40:35.092901 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f852l"] Dec 01 22:40:35 crc kubenswrapper[4857]: I1201 22:40:35.848570 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3448686-7358-45a7-bd12-26d455bea6ff" path="/var/lib/kubelet/pods/b3448686-7358-45a7-bd12-26d455bea6ff/volumes" Dec 01 22:40:35 crc kubenswrapper[4857]: I1201 22:40:35.849737 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b689b746-de83-4849-9a5f-3c368704dced" path="/var/lib/kubelet/pods/b689b746-de83-4849-9a5f-3c368704dced/volumes" Dec 01 22:40:45 crc kubenswrapper[4857]: I1201 22:40:45.835021 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:40:45 crc kubenswrapper[4857]: E1201 22:40:45.835829 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.586190 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wrqrc"] Dec 01 22:40:49 crc kubenswrapper[4857]: E1201 22:40:49.587019 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3448686-7358-45a7-bd12-26d455bea6ff" containerName="extract-content" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.587032 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3448686-7358-45a7-bd12-26d455bea6ff" containerName="extract-content" Dec 01 22:40:49 crc kubenswrapper[4857]: E1201 22:40:49.587056 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3448686-7358-45a7-bd12-26d455bea6ff" containerName="registry-server" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.587062 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3448686-7358-45a7-bd12-26d455bea6ff" containerName="registry-server" Dec 01 22:40:49 crc kubenswrapper[4857]: E1201 22:40:49.587073 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b689b746-de83-4849-9a5f-3c368704dced" containerName="extract-utilities" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.587081 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b689b746-de83-4849-9a5f-3c368704dced" containerName="extract-utilities" Dec 01 22:40:49 crc kubenswrapper[4857]: E1201 22:40:49.587091 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b689b746-de83-4849-9a5f-3c368704dced" containerName="extract-content" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.587096 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b689b746-de83-4849-9a5f-3c368704dced" containerName="extract-content" Dec 01 22:40:49 crc kubenswrapper[4857]: E1201 22:40:49.587112 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b689b746-de83-4849-9a5f-3c368704dced" containerName="registry-server" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.587117 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b689b746-de83-4849-9a5f-3c368704dced" containerName="registry-server" Dec 01 22:40:49 crc kubenswrapper[4857]: E1201 22:40:49.587144 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3448686-7358-45a7-bd12-26d455bea6ff" containerName="extract-utilities" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.587150 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3448686-7358-45a7-bd12-26d455bea6ff" containerName="extract-utilities" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.587317 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b689b746-de83-4849-9a5f-3c368704dced" containerName="registry-server" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.587339 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3448686-7358-45a7-bd12-26d455bea6ff" containerName="registry-server" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.588772 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.615217 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wrqrc"] Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.768706 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4db9473-0f85-4431-b7b1-f3630a76d884-catalog-content\") pod \"redhat-marketplace-wrqrc\" (UID: \"b4db9473-0f85-4431-b7b1-f3630a76d884\") " pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.768896 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4db9473-0f85-4431-b7b1-f3630a76d884-utilities\") pod \"redhat-marketplace-wrqrc\" (UID: \"b4db9473-0f85-4431-b7b1-f3630a76d884\") " pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.768928 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtz4x\" (UniqueName: \"kubernetes.io/projected/b4db9473-0f85-4431-b7b1-f3630a76d884-kube-api-access-rtz4x\") pod \"redhat-marketplace-wrqrc\" (UID: \"b4db9473-0f85-4431-b7b1-f3630a76d884\") " pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.870410 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4db9473-0f85-4431-b7b1-f3630a76d884-catalog-content\") pod \"redhat-marketplace-wrqrc\" (UID: \"b4db9473-0f85-4431-b7b1-f3630a76d884\") " pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.870574 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4db9473-0f85-4431-b7b1-f3630a76d884-utilities\") pod \"redhat-marketplace-wrqrc\" (UID: \"b4db9473-0f85-4431-b7b1-f3630a76d884\") " pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.870593 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtz4x\" (UniqueName: \"kubernetes.io/projected/b4db9473-0f85-4431-b7b1-f3630a76d884-kube-api-access-rtz4x\") pod \"redhat-marketplace-wrqrc\" (UID: \"b4db9473-0f85-4431-b7b1-f3630a76d884\") " pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.871317 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4db9473-0f85-4431-b7b1-f3630a76d884-catalog-content\") pod \"redhat-marketplace-wrqrc\" (UID: \"b4db9473-0f85-4431-b7b1-f3630a76d884\") " pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.871531 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4db9473-0f85-4431-b7b1-f3630a76d884-utilities\") pod \"redhat-marketplace-wrqrc\" (UID: \"b4db9473-0f85-4431-b7b1-f3630a76d884\") " pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.888757 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtz4x\" (UniqueName: \"kubernetes.io/projected/b4db9473-0f85-4431-b7b1-f3630a76d884-kube-api-access-rtz4x\") pod \"redhat-marketplace-wrqrc\" (UID: \"b4db9473-0f85-4431-b7b1-f3630a76d884\") " pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:40:49 crc kubenswrapper[4857]: I1201 22:40:49.912993 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:40:50 crc kubenswrapper[4857]: I1201 22:40:50.491225 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wrqrc"] Dec 01 22:40:50 crc kubenswrapper[4857]: I1201 22:40:50.945827 4857 generic.go:334] "Generic (PLEG): container finished" podID="b4db9473-0f85-4431-b7b1-f3630a76d884" containerID="96dab71e7508cbbcc6b7a5e4b5435f8eb47dde28bd73498bec459040a718635b" exitCode=0 Dec 01 22:40:50 crc kubenswrapper[4857]: I1201 22:40:50.946077 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wrqrc" event={"ID":"b4db9473-0f85-4431-b7b1-f3630a76d884","Type":"ContainerDied","Data":"96dab71e7508cbbcc6b7a5e4b5435f8eb47dde28bd73498bec459040a718635b"} Dec 01 22:40:50 crc kubenswrapper[4857]: I1201 22:40:50.946105 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wrqrc" event={"ID":"b4db9473-0f85-4431-b7b1-f3630a76d884","Type":"ContainerStarted","Data":"466e1507f90093179680b9c90d39ba6c38e1336f207ae3608b0179ba9457e21c"} Dec 01 22:40:52 crc kubenswrapper[4857]: I1201 22:40:52.972437 4857 generic.go:334] "Generic (PLEG): container finished" podID="b4db9473-0f85-4431-b7b1-f3630a76d884" containerID="a64e60719be23537d57fe5d08926f5a29e7df14f9c5142771be10997f18a2001" exitCode=0 Dec 01 22:40:52 crc kubenswrapper[4857]: I1201 22:40:52.972522 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wrqrc" event={"ID":"b4db9473-0f85-4431-b7b1-f3630a76d884","Type":"ContainerDied","Data":"a64e60719be23537d57fe5d08926f5a29e7df14f9c5142771be10997f18a2001"} Dec 01 22:40:53 crc kubenswrapper[4857]: I1201 22:40:53.987125 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wrqrc" event={"ID":"b4db9473-0f85-4431-b7b1-f3630a76d884","Type":"ContainerStarted","Data":"8b2d40e9d8c740048854c7f8b945176e480276c175876206ecb78b9a568d3e73"} Dec 01 22:40:54 crc kubenswrapper[4857]: I1201 22:40:54.014822 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wrqrc" podStartSLOduration=2.445607959 podStartE2EDuration="5.014795134s" podCreationTimestamp="2025-12-01 22:40:49 +0000 UTC" firstStartedPulling="2025-12-01 22:40:50.948637852 +0000 UTC m=+3969.438700169" lastFinishedPulling="2025-12-01 22:40:53.517824987 +0000 UTC m=+3972.007887344" observedRunningTime="2025-12-01 22:40:54.006974544 +0000 UTC m=+3972.497036861" watchObservedRunningTime="2025-12-01 22:40:54.014795134 +0000 UTC m=+3972.504857491" Dec 01 22:40:58 crc kubenswrapper[4857]: I1201 22:40:58.836346 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:40:58 crc kubenswrapper[4857]: E1201 22:40:58.837304 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:40:59 crc kubenswrapper[4857]: I1201 22:40:59.914155 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:40:59 crc kubenswrapper[4857]: I1201 22:40:59.914734 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:41:00 crc kubenswrapper[4857]: I1201 22:41:00.001925 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:41:00 crc kubenswrapper[4857]: I1201 22:41:00.126928 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:41:00 crc kubenswrapper[4857]: I1201 22:41:00.252004 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wrqrc"] Dec 01 22:41:02 crc kubenswrapper[4857]: I1201 22:41:02.077560 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wrqrc" podUID="b4db9473-0f85-4431-b7b1-f3630a76d884" containerName="registry-server" containerID="cri-o://8b2d40e9d8c740048854c7f8b945176e480276c175876206ecb78b9a568d3e73" gracePeriod=2 Dec 01 22:41:02 crc kubenswrapper[4857]: E1201 22:41:02.930813 4857 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4db9473_0f85_4431_b7b1_f3630a76d884.slice/crio-conmon-8b2d40e9d8c740048854c7f8b945176e480276c175876206ecb78b9a568d3e73.scope\": RecentStats: unable to find data in memory cache]" Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.090598 4857 generic.go:334] "Generic (PLEG): container finished" podID="b4db9473-0f85-4431-b7b1-f3630a76d884" containerID="8b2d40e9d8c740048854c7f8b945176e480276c175876206ecb78b9a568d3e73" exitCode=0 Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.090640 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wrqrc" event={"ID":"b4db9473-0f85-4431-b7b1-f3630a76d884","Type":"ContainerDied","Data":"8b2d40e9d8c740048854c7f8b945176e480276c175876206ecb78b9a568d3e73"} Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.090669 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wrqrc" event={"ID":"b4db9473-0f85-4431-b7b1-f3630a76d884","Type":"ContainerDied","Data":"466e1507f90093179680b9c90d39ba6c38e1336f207ae3608b0179ba9457e21c"} Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.090679 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="466e1507f90093179680b9c90d39ba6c38e1336f207ae3608b0179ba9457e21c" Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.184804 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.290537 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4db9473-0f85-4431-b7b1-f3630a76d884-utilities\") pod \"b4db9473-0f85-4431-b7b1-f3630a76d884\" (UID: \"b4db9473-0f85-4431-b7b1-f3630a76d884\") " Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.290856 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtz4x\" (UniqueName: \"kubernetes.io/projected/b4db9473-0f85-4431-b7b1-f3630a76d884-kube-api-access-rtz4x\") pod \"b4db9473-0f85-4431-b7b1-f3630a76d884\" (UID: \"b4db9473-0f85-4431-b7b1-f3630a76d884\") " Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.290953 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4db9473-0f85-4431-b7b1-f3630a76d884-catalog-content\") pod \"b4db9473-0f85-4431-b7b1-f3630a76d884\" (UID: \"b4db9473-0f85-4431-b7b1-f3630a76d884\") " Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.292331 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4db9473-0f85-4431-b7b1-f3630a76d884-utilities" (OuterVolumeSpecName: "utilities") pod "b4db9473-0f85-4431-b7b1-f3630a76d884" (UID: "b4db9473-0f85-4431-b7b1-f3630a76d884"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.303940 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4db9473-0f85-4431-b7b1-f3630a76d884-kube-api-access-rtz4x" (OuterVolumeSpecName: "kube-api-access-rtz4x") pod "b4db9473-0f85-4431-b7b1-f3630a76d884" (UID: "b4db9473-0f85-4431-b7b1-f3630a76d884"). InnerVolumeSpecName "kube-api-access-rtz4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.308426 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4db9473-0f85-4431-b7b1-f3630a76d884-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4db9473-0f85-4431-b7b1-f3630a76d884" (UID: "b4db9473-0f85-4431-b7b1-f3630a76d884"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.393663 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4db9473-0f85-4431-b7b1-f3630a76d884-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.393722 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtz4x\" (UniqueName: \"kubernetes.io/projected/b4db9473-0f85-4431-b7b1-f3630a76d884-kube-api-access-rtz4x\") on node \"crc\" DevicePath \"\"" Dec 01 22:41:03 crc kubenswrapper[4857]: I1201 22:41:03.393746 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4db9473-0f85-4431-b7b1-f3630a76d884-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:41:04 crc kubenswrapper[4857]: I1201 22:41:04.102646 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wrqrc" Dec 01 22:41:04 crc kubenswrapper[4857]: I1201 22:41:04.140163 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wrqrc"] Dec 01 22:41:04 crc kubenswrapper[4857]: I1201 22:41:04.157623 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wrqrc"] Dec 01 22:41:05 crc kubenswrapper[4857]: I1201 22:41:05.848058 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4db9473-0f85-4431-b7b1-f3630a76d884" path="/var/lib/kubelet/pods/b4db9473-0f85-4431-b7b1-f3630a76d884/volumes" Dec 01 22:41:06 crc kubenswrapper[4857]: I1201 22:41:06.374423 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f765fc7d6-x797d_ff230e58-1711-4638-89d6-daf91fb3bfc1/manager/0.log" Dec 01 22:41:12 crc kubenswrapper[4857]: I1201 22:41:12.835145 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:41:12 crc kubenswrapper[4857]: E1201 22:41:12.835805 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:41:23 crc kubenswrapper[4857]: I1201 22:41:23.835627 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:41:23 crc kubenswrapper[4857]: E1201 22:41:23.836792 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.283302 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mgsk4/must-gather-glvmw"] Dec 01 22:41:25 crc kubenswrapper[4857]: E1201 22:41:25.284031 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4db9473-0f85-4431-b7b1-f3630a76d884" containerName="extract-utilities" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.284063 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4db9473-0f85-4431-b7b1-f3630a76d884" containerName="extract-utilities" Dec 01 22:41:25 crc kubenswrapper[4857]: E1201 22:41:25.284097 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4db9473-0f85-4431-b7b1-f3630a76d884" containerName="extract-content" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.284105 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4db9473-0f85-4431-b7b1-f3630a76d884" containerName="extract-content" Dec 01 22:41:25 crc kubenswrapper[4857]: E1201 22:41:25.284122 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4db9473-0f85-4431-b7b1-f3630a76d884" containerName="registry-server" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.284130 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4db9473-0f85-4431-b7b1-f3630a76d884" containerName="registry-server" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.284382 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4db9473-0f85-4431-b7b1-f3630a76d884" containerName="registry-server" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.285712 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgsk4/must-gather-glvmw" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.287956 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mgsk4"/"kube-root-ca.crt" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.289716 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mgsk4"/"openshift-service-ca.crt" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.322604 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mgsk4/must-gather-glvmw"] Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.355813 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jscz\" (UniqueName: \"kubernetes.io/projected/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d-kube-api-access-9jscz\") pod \"must-gather-glvmw\" (UID: \"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d\") " pod="openshift-must-gather-mgsk4/must-gather-glvmw" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.355893 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d-must-gather-output\") pod \"must-gather-glvmw\" (UID: \"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d\") " pod="openshift-must-gather-mgsk4/must-gather-glvmw" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.457497 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jscz\" (UniqueName: \"kubernetes.io/projected/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d-kube-api-access-9jscz\") pod \"must-gather-glvmw\" (UID: \"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d\") " pod="openshift-must-gather-mgsk4/must-gather-glvmw" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.457609 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d-must-gather-output\") pod \"must-gather-glvmw\" (UID: \"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d\") " pod="openshift-must-gather-mgsk4/must-gather-glvmw" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.458169 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d-must-gather-output\") pod \"must-gather-glvmw\" (UID: \"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d\") " pod="openshift-must-gather-mgsk4/must-gather-glvmw" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.477583 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jscz\" (UniqueName: \"kubernetes.io/projected/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d-kube-api-access-9jscz\") pod \"must-gather-glvmw\" (UID: \"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d\") " pod="openshift-must-gather-mgsk4/must-gather-glvmw" Dec 01 22:41:25 crc kubenswrapper[4857]: I1201 22:41:25.602537 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgsk4/must-gather-glvmw" Dec 01 22:41:26 crc kubenswrapper[4857]: I1201 22:41:26.144854 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mgsk4/must-gather-glvmw"] Dec 01 22:41:26 crc kubenswrapper[4857]: I1201 22:41:26.368834 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgsk4/must-gather-glvmw" event={"ID":"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d","Type":"ContainerStarted","Data":"a3c2a4d1ba06eab6db5604d7b18b6b40e7d2cad3743e12513602234b7369d4b7"} Dec 01 22:41:31 crc kubenswrapper[4857]: I1201 22:41:31.441962 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgsk4/must-gather-glvmw" event={"ID":"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d","Type":"ContainerStarted","Data":"8371467cd8fd4561b871b544cd3c878f9dfc722c5ecfab0507e19e49ecc732e7"} Dec 01 22:41:31 crc kubenswrapper[4857]: I1201 22:41:31.442499 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgsk4/must-gather-glvmw" event={"ID":"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d","Type":"ContainerStarted","Data":"27e63ca2383fb545df3332964fe2e89e05b70c3cfa909a0777cc96913abc4fa8"} Dec 01 22:41:31 crc kubenswrapper[4857]: I1201 22:41:31.476751 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mgsk4/must-gather-glvmw" podStartSLOduration=2.335892849 podStartE2EDuration="6.476722208s" podCreationTimestamp="2025-12-01 22:41:25 +0000 UTC" firstStartedPulling="2025-12-01 22:41:26.136653265 +0000 UTC m=+4004.626715592" lastFinishedPulling="2025-12-01 22:41:30.277482594 +0000 UTC m=+4008.767544951" observedRunningTime="2025-12-01 22:41:31.462598445 +0000 UTC m=+4009.952660762" watchObservedRunningTime="2025-12-01 22:41:31.476722208 +0000 UTC m=+4009.966784565" Dec 01 22:41:34 crc kubenswrapper[4857]: I1201 22:41:34.834881 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:41:34 crc kubenswrapper[4857]: E1201 22:41:34.835634 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:41:35 crc kubenswrapper[4857]: I1201 22:41:35.501619 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mgsk4/crc-debug-5nq5p"] Dec 01 22:41:35 crc kubenswrapper[4857]: I1201 22:41:35.503800 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" Dec 01 22:41:35 crc kubenswrapper[4857]: I1201 22:41:35.508673 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mgsk4"/"default-dockercfg-r7ltw" Dec 01 22:41:35 crc kubenswrapper[4857]: I1201 22:41:35.625346 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmn6q\" (UniqueName: \"kubernetes.io/projected/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1-kube-api-access-dmn6q\") pod \"crc-debug-5nq5p\" (UID: \"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1\") " pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" Dec 01 22:41:35 crc kubenswrapper[4857]: I1201 22:41:35.625614 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1-host\") pod \"crc-debug-5nq5p\" (UID: \"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1\") " pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" Dec 01 22:41:35 crc kubenswrapper[4857]: I1201 22:41:35.727448 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1-host\") pod \"crc-debug-5nq5p\" (UID: \"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1\") " pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" Dec 01 22:41:35 crc kubenswrapper[4857]: I1201 22:41:35.727587 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmn6q\" (UniqueName: \"kubernetes.io/projected/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1-kube-api-access-dmn6q\") pod \"crc-debug-5nq5p\" (UID: \"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1\") " pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" Dec 01 22:41:35 crc kubenswrapper[4857]: I1201 22:41:35.727639 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1-host\") pod \"crc-debug-5nq5p\" (UID: \"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1\") " pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" Dec 01 22:41:35 crc kubenswrapper[4857]: I1201 22:41:35.760827 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmn6q\" (UniqueName: \"kubernetes.io/projected/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1-kube-api-access-dmn6q\") pod \"crc-debug-5nq5p\" (UID: \"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1\") " pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" Dec 01 22:41:35 crc kubenswrapper[4857]: I1201 22:41:35.823111 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" Dec 01 22:41:35 crc kubenswrapper[4857]: W1201 22:41:35.875992 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc19f9ce_6f54_4ea9_a5b6_d10db71103f1.slice/crio-ecd96cb6f21ece7792cc85e16766854414851be7fc936ad9f73289abdf7394ca WatchSource:0}: Error finding container ecd96cb6f21ece7792cc85e16766854414851be7fc936ad9f73289abdf7394ca: Status 404 returned error can't find the container with id ecd96cb6f21ece7792cc85e16766854414851be7fc936ad9f73289abdf7394ca Dec 01 22:41:36 crc kubenswrapper[4857]: I1201 22:41:36.515417 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" event={"ID":"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1","Type":"ContainerStarted","Data":"ecd96cb6f21ece7792cc85e16766854414851be7fc936ad9f73289abdf7394ca"} Dec 01 22:41:46 crc kubenswrapper[4857]: I1201 22:41:46.834991 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:41:46 crc kubenswrapper[4857]: E1201 22:41:46.835879 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:41:47 crc kubenswrapper[4857]: I1201 22:41:47.624330 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" event={"ID":"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1","Type":"ContainerStarted","Data":"d362ad4cec62e9907637ca86493a21bc47c59a19dda83e36c36a13efccb742ac"} Dec 01 22:41:47 crc kubenswrapper[4857]: I1201 22:41:47.647986 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" podStartSLOduration=1.154025204 podStartE2EDuration="12.647963745s" podCreationTimestamp="2025-12-01 22:41:35 +0000 UTC" firstStartedPulling="2025-12-01 22:41:35.87898018 +0000 UTC m=+4014.369042507" lastFinishedPulling="2025-12-01 22:41:47.372918711 +0000 UTC m=+4025.862981048" observedRunningTime="2025-12-01 22:41:47.644348397 +0000 UTC m=+4026.134410714" watchObservedRunningTime="2025-12-01 22:41:47.647963745 +0000 UTC m=+4026.138026072" Dec 01 22:41:57 crc kubenswrapper[4857]: I1201 22:41:57.834958 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:41:57 crc kubenswrapper[4857]: E1201 22:41:57.835969 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:42:03 crc kubenswrapper[4857]: I1201 22:42:03.786347 4857 generic.go:334] "Generic (PLEG): container finished" podID="bc19f9ce-6f54-4ea9-a5b6-d10db71103f1" containerID="d362ad4cec62e9907637ca86493a21bc47c59a19dda83e36c36a13efccb742ac" exitCode=0 Dec 01 22:42:03 crc kubenswrapper[4857]: I1201 22:42:03.786519 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" event={"ID":"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1","Type":"ContainerDied","Data":"d362ad4cec62e9907637ca86493a21bc47c59a19dda83e36c36a13efccb742ac"} Dec 01 22:42:04 crc kubenswrapper[4857]: I1201 22:42:04.931812 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" Dec 01 22:42:04 crc kubenswrapper[4857]: I1201 22:42:04.974402 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mgsk4/crc-debug-5nq5p"] Dec 01 22:42:04 crc kubenswrapper[4857]: I1201 22:42:04.984871 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mgsk4/crc-debug-5nq5p"] Dec 01 22:42:05 crc kubenswrapper[4857]: I1201 22:42:05.064910 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmn6q\" (UniqueName: \"kubernetes.io/projected/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1-kube-api-access-dmn6q\") pod \"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1\" (UID: \"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1\") " Dec 01 22:42:05 crc kubenswrapper[4857]: I1201 22:42:05.065285 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1-host\") pod \"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1\" (UID: \"bc19f9ce-6f54-4ea9-a5b6-d10db71103f1\") " Dec 01 22:42:05 crc kubenswrapper[4857]: I1201 22:42:05.065416 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1-host" (OuterVolumeSpecName: "host") pod "bc19f9ce-6f54-4ea9-a5b6-d10db71103f1" (UID: "bc19f9ce-6f54-4ea9-a5b6-d10db71103f1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 22:42:05 crc kubenswrapper[4857]: I1201 22:42:05.065853 4857 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1-host\") on node \"crc\" DevicePath \"\"" Dec 01 22:42:05 crc kubenswrapper[4857]: I1201 22:42:05.075268 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1-kube-api-access-dmn6q" (OuterVolumeSpecName: "kube-api-access-dmn6q") pod "bc19f9ce-6f54-4ea9-a5b6-d10db71103f1" (UID: "bc19f9ce-6f54-4ea9-a5b6-d10db71103f1"). InnerVolumeSpecName "kube-api-access-dmn6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:42:05 crc kubenswrapper[4857]: I1201 22:42:05.166762 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmn6q\" (UniqueName: \"kubernetes.io/projected/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1-kube-api-access-dmn6q\") on node \"crc\" DevicePath \"\"" Dec 01 22:42:05 crc kubenswrapper[4857]: I1201 22:42:05.822257 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecd96cb6f21ece7792cc85e16766854414851be7fc936ad9f73289abdf7394ca" Dec 01 22:42:05 crc kubenswrapper[4857]: I1201 22:42:05.822505 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgsk4/crc-debug-5nq5p" Dec 01 22:42:05 crc kubenswrapper[4857]: I1201 22:42:05.848824 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc19f9ce-6f54-4ea9-a5b6-d10db71103f1" path="/var/lib/kubelet/pods/bc19f9ce-6f54-4ea9-a5b6-d10db71103f1/volumes" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.197389 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mgsk4/crc-debug-92qbc"] Dec 01 22:42:06 crc kubenswrapper[4857]: E1201 22:42:06.197998 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc19f9ce-6f54-4ea9-a5b6-d10db71103f1" containerName="container-00" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.198018 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc19f9ce-6f54-4ea9-a5b6-d10db71103f1" containerName="container-00" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.198416 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc19f9ce-6f54-4ea9-a5b6-d10db71103f1" containerName="container-00" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.203209 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgsk4/crc-debug-92qbc" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.207453 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mgsk4"/"default-dockercfg-r7ltw" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.389778 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9a5cc2c-8b4f-412b-8eaa-90084d966507-host\") pod \"crc-debug-92qbc\" (UID: \"c9a5cc2c-8b4f-412b-8eaa-90084d966507\") " pod="openshift-must-gather-mgsk4/crc-debug-92qbc" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.389926 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k8v8\" (UniqueName: \"kubernetes.io/projected/c9a5cc2c-8b4f-412b-8eaa-90084d966507-kube-api-access-5k8v8\") pod \"crc-debug-92qbc\" (UID: \"c9a5cc2c-8b4f-412b-8eaa-90084d966507\") " pod="openshift-must-gather-mgsk4/crc-debug-92qbc" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.492191 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9a5cc2c-8b4f-412b-8eaa-90084d966507-host\") pod \"crc-debug-92qbc\" (UID: \"c9a5cc2c-8b4f-412b-8eaa-90084d966507\") " pod="openshift-must-gather-mgsk4/crc-debug-92qbc" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.492342 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9a5cc2c-8b4f-412b-8eaa-90084d966507-host\") pod \"crc-debug-92qbc\" (UID: \"c9a5cc2c-8b4f-412b-8eaa-90084d966507\") " pod="openshift-must-gather-mgsk4/crc-debug-92qbc" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.492471 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k8v8\" (UniqueName: \"kubernetes.io/projected/c9a5cc2c-8b4f-412b-8eaa-90084d966507-kube-api-access-5k8v8\") pod \"crc-debug-92qbc\" (UID: \"c9a5cc2c-8b4f-412b-8eaa-90084d966507\") " pod="openshift-must-gather-mgsk4/crc-debug-92qbc" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.512893 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k8v8\" (UniqueName: \"kubernetes.io/projected/c9a5cc2c-8b4f-412b-8eaa-90084d966507-kube-api-access-5k8v8\") pod \"crc-debug-92qbc\" (UID: \"c9a5cc2c-8b4f-412b-8eaa-90084d966507\") " pod="openshift-must-gather-mgsk4/crc-debug-92qbc" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.526782 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgsk4/crc-debug-92qbc" Dec 01 22:42:06 crc kubenswrapper[4857]: I1201 22:42:06.833234 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgsk4/crc-debug-92qbc" event={"ID":"c9a5cc2c-8b4f-412b-8eaa-90084d966507","Type":"ContainerStarted","Data":"f8e444c1a681347833ae4659f2f82c934f1e38a094365dd9edafa818c8ded59d"} Dec 01 22:42:07 crc kubenswrapper[4857]: I1201 22:42:07.845937 4857 generic.go:334] "Generic (PLEG): container finished" podID="c9a5cc2c-8b4f-412b-8eaa-90084d966507" containerID="584f26187625f5b8ce6df56d620bf862f1199eed90c8e2579b80b9dbdf932e5a" exitCode=1 Dec 01 22:42:07 crc kubenswrapper[4857]: I1201 22:42:07.849215 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgsk4/crc-debug-92qbc" event={"ID":"c9a5cc2c-8b4f-412b-8eaa-90084d966507","Type":"ContainerDied","Data":"584f26187625f5b8ce6df56d620bf862f1199eed90c8e2579b80b9dbdf932e5a"} Dec 01 22:42:07 crc kubenswrapper[4857]: I1201 22:42:07.895417 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mgsk4/crc-debug-92qbc"] Dec 01 22:42:07 crc kubenswrapper[4857]: I1201 22:42:07.908349 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mgsk4/crc-debug-92qbc"] Dec 01 22:42:08 crc kubenswrapper[4857]: I1201 22:42:08.959939 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgsk4/crc-debug-92qbc" Dec 01 22:42:09 crc kubenswrapper[4857]: I1201 22:42:09.079695 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k8v8\" (UniqueName: \"kubernetes.io/projected/c9a5cc2c-8b4f-412b-8eaa-90084d966507-kube-api-access-5k8v8\") pod \"c9a5cc2c-8b4f-412b-8eaa-90084d966507\" (UID: \"c9a5cc2c-8b4f-412b-8eaa-90084d966507\") " Dec 01 22:42:09 crc kubenswrapper[4857]: I1201 22:42:09.079815 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9a5cc2c-8b4f-412b-8eaa-90084d966507-host\") pod \"c9a5cc2c-8b4f-412b-8eaa-90084d966507\" (UID: \"c9a5cc2c-8b4f-412b-8eaa-90084d966507\") " Dec 01 22:42:09 crc kubenswrapper[4857]: I1201 22:42:09.080117 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c9a5cc2c-8b4f-412b-8eaa-90084d966507-host" (OuterVolumeSpecName: "host") pod "c9a5cc2c-8b4f-412b-8eaa-90084d966507" (UID: "c9a5cc2c-8b4f-412b-8eaa-90084d966507"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 22:42:09 crc kubenswrapper[4857]: I1201 22:42:09.080650 4857 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9a5cc2c-8b4f-412b-8eaa-90084d966507-host\") on node \"crc\" DevicePath \"\"" Dec 01 22:42:09 crc kubenswrapper[4857]: I1201 22:42:09.090555 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9a5cc2c-8b4f-412b-8eaa-90084d966507-kube-api-access-5k8v8" (OuterVolumeSpecName: "kube-api-access-5k8v8") pod "c9a5cc2c-8b4f-412b-8eaa-90084d966507" (UID: "c9a5cc2c-8b4f-412b-8eaa-90084d966507"). InnerVolumeSpecName "kube-api-access-5k8v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:42:09 crc kubenswrapper[4857]: I1201 22:42:09.182347 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k8v8\" (UniqueName: \"kubernetes.io/projected/c9a5cc2c-8b4f-412b-8eaa-90084d966507-kube-api-access-5k8v8\") on node \"crc\" DevicePath \"\"" Dec 01 22:42:09 crc kubenswrapper[4857]: I1201 22:42:09.838765 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:42:09 crc kubenswrapper[4857]: E1201 22:42:09.839083 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:42:09 crc kubenswrapper[4857]: I1201 22:42:09.848669 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9a5cc2c-8b4f-412b-8eaa-90084d966507" path="/var/lib/kubelet/pods/c9a5cc2c-8b4f-412b-8eaa-90084d966507/volumes" Dec 01 22:42:09 crc kubenswrapper[4857]: I1201 22:42:09.865640 4857 scope.go:117] "RemoveContainer" containerID="584f26187625f5b8ce6df56d620bf862f1199eed90c8e2579b80b9dbdf932e5a" Dec 01 22:42:09 crc kubenswrapper[4857]: I1201 22:42:09.865793 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgsk4/crc-debug-92qbc" Dec 01 22:42:22 crc kubenswrapper[4857]: I1201 22:42:22.835396 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:42:22 crc kubenswrapper[4857]: E1201 22:42:22.836458 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:42:34 crc kubenswrapper[4857]: I1201 22:42:34.835481 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:42:34 crc kubenswrapper[4857]: E1201 22:42:34.836577 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:42:47 crc kubenswrapper[4857]: I1201 22:42:47.835431 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:42:47 crc kubenswrapper[4857]: E1201 22:42:47.836187 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:43:00 crc kubenswrapper[4857]: I1201 22:43:00.145065 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_98dbb0ff-4644-431e-801e-5ce7fb3f6f81/init-config-reloader/0.log" Dec 01 22:43:00 crc kubenswrapper[4857]: I1201 22:43:00.408379 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_98dbb0ff-4644-431e-801e-5ce7fb3f6f81/alertmanager/0.log" Dec 01 22:43:00 crc kubenswrapper[4857]: I1201 22:43:00.436261 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_98dbb0ff-4644-431e-801e-5ce7fb3f6f81/init-config-reloader/0.log" Dec 01 22:43:00 crc kubenswrapper[4857]: I1201 22:43:00.458772 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_98dbb0ff-4644-431e-801e-5ce7fb3f6f81/config-reloader/0.log" Dec 01 22:43:00 crc kubenswrapper[4857]: I1201 22:43:00.641910 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e78cc157-06d0-4c5f-8da8-6e1a9591600a/aodh-evaluator/0.log" Dec 01 22:43:00 crc kubenswrapper[4857]: I1201 22:43:00.646533 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e78cc157-06d0-4c5f-8da8-6e1a9591600a/aodh-api/0.log" Dec 01 22:43:00 crc kubenswrapper[4857]: I1201 22:43:00.659867 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e78cc157-06d0-4c5f-8da8-6e1a9591600a/aodh-listener/0.log" Dec 01 22:43:00 crc kubenswrapper[4857]: I1201 22:43:00.832180 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e78cc157-06d0-4c5f-8da8-6e1a9591600a/aodh-notifier/0.log" Dec 01 22:43:00 crc kubenswrapper[4857]: I1201 22:43:00.873494 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-cf44fcf6b-c86cx_13134eb9-c4d1-4060-8fd3-266afd46f420/barbican-api/0.log" Dec 01 22:43:00 crc kubenswrapper[4857]: I1201 22:43:00.888780 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-cf44fcf6b-c86cx_13134eb9-c4d1-4060-8fd3-266afd46f420/barbican-api-log/0.log" Dec 01 22:43:01 crc kubenswrapper[4857]: I1201 22:43:01.351508 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-db75d4bb-tjp4b_d92f4970-9829-4417-9cc1-9436c8993a18/barbican-keystone-listener/0.log" Dec 01 22:43:01 crc kubenswrapper[4857]: I1201 22:43:01.440320 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-db75d4bb-tjp4b_d92f4970-9829-4417-9cc1-9436c8993a18/barbican-keystone-listener-log/0.log" Dec 01 22:43:01 crc kubenswrapper[4857]: I1201 22:43:01.543478 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5f6bc6688c-77x7p_5ae783f5-078d-49c0-9b57-cf796b0e8f3b/barbican-worker-log/0.log" Dec 01 22:43:01 crc kubenswrapper[4857]: I1201 22:43:01.547272 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5f6bc6688c-77x7p_5ae783f5-078d-49c0-9b57-cf796b0e8f3b/barbican-worker/0.log" Dec 01 22:43:01 crc kubenswrapper[4857]: I1201 22:43:01.694601 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt_3916d4da-634b-4ade-be14-3a0c60e6a9d2/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:01 crc kubenswrapper[4857]: I1201 22:43:01.788275 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ccfb3129-79e9-4397-93c3-05ddbd0e6f76/ceilometer-central-agent/0.log" Dec 01 22:43:01 crc kubenswrapper[4857]: I1201 22:43:01.879827 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ccfb3129-79e9-4397-93c3-05ddbd0e6f76/ceilometer-notification-agent/0.log" Dec 01 22:43:01 crc kubenswrapper[4857]: I1201 22:43:01.905253 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ccfb3129-79e9-4397-93c3-05ddbd0e6f76/proxy-httpd/0.log" Dec 01 22:43:01 crc kubenswrapper[4857]: I1201 22:43:01.993677 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ccfb3129-79e9-4397-93c3-05ddbd0e6f76/sg-core/0.log" Dec 01 22:43:02 crc kubenswrapper[4857]: I1201 22:43:02.095827 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_919f0ad8-05ea-44a1-a4ae-f81f4255e8da/cinder-api/0.log" Dec 01 22:43:02 crc kubenswrapper[4857]: I1201 22:43:02.110153 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_919f0ad8-05ea-44a1-a4ae-f81f4255e8da/cinder-api-log/0.log" Dec 01 22:43:02 crc kubenswrapper[4857]: I1201 22:43:02.322953 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f38eac29-d14a-4bf8-bbd8-df39c591ee6a/cinder-scheduler/0.log" Dec 01 22:43:02 crc kubenswrapper[4857]: I1201 22:43:02.531453 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f38eac29-d14a-4bf8-bbd8-df39c591ee6a/probe/0.log" Dec 01 22:43:02 crc kubenswrapper[4857]: I1201 22:43:02.681350 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm_0a768b62-44a4-4493-bd0d-2a536c55370d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:02 crc kubenswrapper[4857]: I1201 22:43:02.743054 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-7mccz_e3196a47-b0ed-4f5d-a639-496ff78d1acd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:02 crc kubenswrapper[4857]: I1201 22:43:02.835524 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:43:02 crc kubenswrapper[4857]: E1201 22:43:02.835833 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:43:02 crc kubenswrapper[4857]: I1201 22:43:02.903342 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-5mp7g_72ba8edd-5f69-4c93-9954-aaa52e18c2fa/init/0.log" Dec 01 22:43:03 crc kubenswrapper[4857]: I1201 22:43:03.092085 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-5mp7g_72ba8edd-5f69-4c93-9954-aaa52e18c2fa/dnsmasq-dns/0.log" Dec 01 22:43:03 crc kubenswrapper[4857]: I1201 22:43:03.123878 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-5mp7g_72ba8edd-5f69-4c93-9954-aaa52e18c2fa/init/0.log" Dec 01 22:43:03 crc kubenswrapper[4857]: I1201 22:43:03.144584 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2_d419ce3b-db4a-4224-a322-69aaa24650fa/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:03 crc kubenswrapper[4857]: I1201 22:43:03.488508 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_dc169c0b-67f2-48bf-8ac6-f58ca383ae4c/glance-httpd/0.log" Dec 01 22:43:03 crc kubenswrapper[4857]: I1201 22:43:03.526086 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_dc169c0b-67f2-48bf-8ac6-f58ca383ae4c/glance-log/0.log" Dec 01 22:43:03 crc kubenswrapper[4857]: I1201 22:43:03.671551 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4f3b7eba-7e44-4f50-a2da-626cfe8e3921/glance-log/0.log" Dec 01 22:43:03 crc kubenswrapper[4857]: I1201 22:43:03.700785 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4f3b7eba-7e44-4f50-a2da-626cfe8e3921/glance-httpd/0.log" Dec 01 22:43:04 crc kubenswrapper[4857]: I1201 22:43:04.027967 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-5c86f5f679-g6dgl_05d88f39-b756-4b31-9385-99213e839003/heat-api/0.log" Dec 01 22:43:04 crc kubenswrapper[4857]: I1201 22:43:04.182010 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-7c64685c4b-rgnfp_d9fd2d2f-01df-4aa4-bc83-46f30fb0c592/heat-engine/0.log" Dec 01 22:43:04 crc kubenswrapper[4857]: I1201 22:43:04.210298 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5894b69764-fpz5r_ce4b4968-ffc0-4d0d-86b5-da72a0712630/heat-cfnapi/0.log" Dec 01 22:43:04 crc kubenswrapper[4857]: I1201 22:43:04.254860 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp_f71f02cd-2132-437a-b18c-e3513eb0710c/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:04 crc kubenswrapper[4857]: I1201 22:43:04.410084 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-2zzdx_696c86fa-0abd-4bae-9821-f570a70ad6e0/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:04 crc kubenswrapper[4857]: I1201 22:43:04.439257 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29410441-m46zq_a8480d29-0c86-4ee6-85ea-6be3087135ec/keystone-cron/0.log" Dec 01 22:43:04 crc kubenswrapper[4857]: I1201 22:43:04.662685 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_cfbeeaca-f2fe-41ed-bb4d-092cb86751d5/kube-state-metrics/0.log" Dec 01 22:43:04 crc kubenswrapper[4857]: I1201 22:43:04.704311 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-fdd4494c7-87llf_085e66d4-f84a-47e2-851b-99f08ccc9683/keystone-api/0.log" Dec 01 22:43:04 crc kubenswrapper[4857]: I1201 22:43:04.834338 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4_afe24796-7855-411d-b62a-a4e5fefc12c6/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:04 crc kubenswrapper[4857]: I1201 22:43:04.994079 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57b8cf5fc7-kg4kw_33d2facf-f0c8-4566-a15c-038f0f9f9a6e/neutron-api/0.log" Dec 01 22:43:05 crc kubenswrapper[4857]: I1201 22:43:05.078767 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57b8cf5fc7-kg4kw_33d2facf-f0c8-4566-a15c-038f0f9f9a6e/neutron-httpd/0.log" Dec 01 22:43:05 crc kubenswrapper[4857]: I1201 22:43:05.226477 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x_60f230d9-9293-4ab7-a2c5-5a9e3c5dd623/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:05 crc kubenswrapper[4857]: I1201 22:43:05.552070 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_dcd76a72-5869-4014-8bf7-ee57297feb85/nova-api-log/0.log" Dec 01 22:43:05 crc kubenswrapper[4857]: I1201 22:43:05.600311 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a69a1083-f915-4820-82f3-f82c5053d365/nova-cell0-conductor-conductor/0.log" Dec 01 22:43:05 crc kubenswrapper[4857]: I1201 22:43:05.748320 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_dcd76a72-5869-4014-8bf7-ee57297feb85/nova-api-api/0.log" Dec 01 22:43:05 crc kubenswrapper[4857]: I1201 22:43:05.818607 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_cd0a02ec-0eec-4459-9de2-96f044ff45c7/nova-cell1-conductor-conductor/0.log" Dec 01 22:43:05 crc kubenswrapper[4857]: I1201 22:43:05.941256 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e28cb1b9-5dff-4741-ac99-f72b10c0c09d/nova-cell1-novncproxy-novncproxy/0.log" Dec 01 22:43:06 crc kubenswrapper[4857]: I1201 22:43:06.105763 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-l7ltc_3b8d39f4-2804-4fda-bec9-0cd4c2401f42/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:06 crc kubenswrapper[4857]: I1201 22:43:06.255192 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f3a844a9-6868-43b6-b5a9-402e619f0d50/nova-metadata-log/0.log" Dec 01 22:43:06 crc kubenswrapper[4857]: I1201 22:43:06.402131 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_2859495a-e026-400f-a304-05e25e961f2e/nova-scheduler-scheduler/0.log" Dec 01 22:43:06 crc kubenswrapper[4857]: I1201 22:43:06.436296 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_be0a242c-3c02-4d1e-910b-21e74e7c4e43/mysql-bootstrap/0.log" Dec 01 22:43:06 crc kubenswrapper[4857]: I1201 22:43:06.763634 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_be0a242c-3c02-4d1e-910b-21e74e7c4e43/mysql-bootstrap/0.log" Dec 01 22:43:06 crc kubenswrapper[4857]: I1201 22:43:06.801293 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_be0a242c-3c02-4d1e-910b-21e74e7c4e43/galera/0.log" Dec 01 22:43:06 crc kubenswrapper[4857]: I1201 22:43:06.926944 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b3755be2-a089-4259-af43-6e0fbedef24b/mysql-bootstrap/0.log" Dec 01 22:43:07 crc kubenswrapper[4857]: I1201 22:43:07.143643 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b3755be2-a089-4259-af43-6e0fbedef24b/mysql-bootstrap/0.log" Dec 01 22:43:07 crc kubenswrapper[4857]: I1201 22:43:07.156968 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b3755be2-a089-4259-af43-6e0fbedef24b/galera/0.log" Dec 01 22:43:07 crc kubenswrapper[4857]: I1201 22:43:07.317670 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_8b715e70-6964-4dd3-9c88-67d41fd7ad7a/openstackclient/0.log" Dec 01 22:43:07 crc kubenswrapper[4857]: I1201 22:43:07.351082 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lqjrc_7aab234d-a569-48bb-b75b-6a4ecf993df7/openstack-network-exporter/0.log" Dec 01 22:43:07 crc kubenswrapper[4857]: I1201 22:43:07.521992 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-nhfhn_32213ad3-4287-4968-9d42-bfeb2ff558d1/ovn-controller/0.log" Dec 01 22:43:07 crc kubenswrapper[4857]: I1201 22:43:07.609741 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f3a844a9-6868-43b6-b5a9-402e619f0d50/nova-metadata-metadata/0.log" Dec 01 22:43:07 crc kubenswrapper[4857]: I1201 22:43:07.756345 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gwbmr_5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f/ovsdb-server-init/0.log" Dec 01 22:43:07 crc kubenswrapper[4857]: I1201 22:43:07.870814 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gwbmr_5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f/ovs-vswitchd/0.log" Dec 01 22:43:07 crc kubenswrapper[4857]: I1201 22:43:07.930499 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gwbmr_5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f/ovsdb-server-init/0.log" Dec 01 22:43:07 crc kubenswrapper[4857]: I1201 22:43:07.978917 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gwbmr_5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f/ovsdb-server/0.log" Dec 01 22:43:08 crc kubenswrapper[4857]: I1201 22:43:08.120963 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-h6nrk_6640e334-5d6f-441b-b213-985398e67768/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:08 crc kubenswrapper[4857]: I1201 22:43:08.205329 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_467e700f-9a49-4c3d-8108-9bb78cc31495/openstack-network-exporter/0.log" Dec 01 22:43:08 crc kubenswrapper[4857]: I1201 22:43:08.238984 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_467e700f-9a49-4c3d-8108-9bb78cc31495/ovn-northd/0.log" Dec 01 22:43:08 crc kubenswrapper[4857]: I1201 22:43:08.400081 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b726f947-1deb-4268-ae76-900394324ecf/openstack-network-exporter/0.log" Dec 01 22:43:08 crc kubenswrapper[4857]: I1201 22:43:08.406791 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b726f947-1deb-4268-ae76-900394324ecf/ovsdbserver-nb/0.log" Dec 01 22:43:08 crc kubenswrapper[4857]: I1201 22:43:08.710513 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3/ovsdbserver-sb/0.log" Dec 01 22:43:08 crc kubenswrapper[4857]: I1201 22:43:08.734695 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3/openstack-network-exporter/0.log" Dec 01 22:43:08 crc kubenswrapper[4857]: I1201 22:43:08.860100 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6965455568-h4s4z_338abaa2-77aa-403c-b26b-6a5d8c02ff9f/placement-api/0.log" Dec 01 22:43:08 crc kubenswrapper[4857]: I1201 22:43:08.943542 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6965455568-h4s4z_338abaa2-77aa-403c-b26b-6a5d8c02ff9f/placement-log/0.log" Dec 01 22:43:09 crc kubenswrapper[4857]: I1201 22:43:09.005215 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0203f043-2b36-4611-9b70-1fd54f97a1e3/init-config-reloader/0.log" Dec 01 22:43:09 crc kubenswrapper[4857]: I1201 22:43:09.171444 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0203f043-2b36-4611-9b70-1fd54f97a1e3/init-config-reloader/0.log" Dec 01 22:43:09 crc kubenswrapper[4857]: I1201 22:43:09.206821 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0203f043-2b36-4611-9b70-1fd54f97a1e3/prometheus/0.log" Dec 01 22:43:09 crc kubenswrapper[4857]: I1201 22:43:09.226742 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0203f043-2b36-4611-9b70-1fd54f97a1e3/config-reloader/0.log" Dec 01 22:43:09 crc kubenswrapper[4857]: I1201 22:43:09.276470 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0203f043-2b36-4611-9b70-1fd54f97a1e3/thanos-sidecar/0.log" Dec 01 22:43:09 crc kubenswrapper[4857]: I1201 22:43:09.429911 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825/setup-container/0.log" Dec 01 22:43:09 crc kubenswrapper[4857]: I1201 22:43:09.671721 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825/rabbitmq/0.log" Dec 01 22:43:09 crc kubenswrapper[4857]: I1201 22:43:09.765122 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a49ab0a8-09f6-4bfa-9276-c89606c0030e/setup-container/0.log" Dec 01 22:43:09 crc kubenswrapper[4857]: I1201 22:43:09.884752 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825/setup-container/0.log" Dec 01 22:43:10 crc kubenswrapper[4857]: I1201 22:43:10.044088 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a49ab0a8-09f6-4bfa-9276-c89606c0030e/setup-container/0.log" Dec 01 22:43:10 crc kubenswrapper[4857]: I1201 22:43:10.067589 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a49ab0a8-09f6-4bfa-9276-c89606c0030e/rabbitmq/0.log" Dec 01 22:43:10 crc kubenswrapper[4857]: I1201 22:43:10.108817 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4_ad6a5aa6-3dc8-4387-9834-7681a0bf77ce/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:10 crc kubenswrapper[4857]: I1201 22:43:10.275151 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-twm8k_8ae0c086-76b5-44dc-ab9f-e8988b92cc02/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:10 crc kubenswrapper[4857]: I1201 22:43:10.330961 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h_bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:10 crc kubenswrapper[4857]: I1201 22:43:10.558420 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-cwc7h_a1548cec-634a-469e-b3bf-1813dfe676c1/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:10 crc kubenswrapper[4857]: I1201 22:43:10.633610 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8zsck_e7a9f9f9-0f72-46cc-9f30-461fe1a33903/ssh-known-hosts-edpm-deployment/0.log" Dec 01 22:43:10 crc kubenswrapper[4857]: I1201 22:43:10.851180 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b8476f857-zrr4q_b43470bb-ee08-4170-94c6-4a3acbbce50e/proxy-server/0.log" Dec 01 22:43:10 crc kubenswrapper[4857]: I1201 22:43:10.981075 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b8476f857-zrr4q_b43470bb-ee08-4170-94c6-4a3acbbce50e/proxy-httpd/0.log" Dec 01 22:43:10 crc kubenswrapper[4857]: I1201 22:43:10.986331 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-q5szn_34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b/swift-ring-rebalance/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.094598 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/account-auditor/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.208799 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/account-replicator/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.215304 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/account-reaper/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.317734 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/account-server/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.372115 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/container-auditor/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.432916 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/container-server/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.444629 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/container-replicator/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.537686 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/container-updater/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.629905 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/object-auditor/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.653518 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/object-expirer/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.662895 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/object-replicator/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.776218 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/object-server/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.861347 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/object-updater/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.893691 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/rsync/0.log" Dec 01 22:43:11 crc kubenswrapper[4857]: I1201 22:43:11.969823 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/swift-recon-cron/0.log" Dec 01 22:43:12 crc kubenswrapper[4857]: I1201 22:43:12.157415 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-6wnms_2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:12 crc kubenswrapper[4857]: I1201 22:43:12.194651 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9_e74258df-2d15-4928-993f-fac533426f97/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:43:15 crc kubenswrapper[4857]: I1201 22:43:15.834691 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:43:15 crc kubenswrapper[4857]: E1201 22:43:15.835399 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:43:19 crc kubenswrapper[4857]: I1201 22:43:19.783262 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_6a290cd2-c321-4cd5-b168-4866c03b135b/memcached/0.log" Dec 01 22:43:26 crc kubenswrapper[4857]: I1201 22:43:26.835434 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:43:26 crc kubenswrapper[4857]: E1201 22:43:26.836177 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:43:37 crc kubenswrapper[4857]: I1201 22:43:37.835563 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:43:37 crc kubenswrapper[4857]: E1201 22:43:37.836344 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:43:39 crc kubenswrapper[4857]: I1201 22:43:39.303603 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/util/0.log" Dec 01 22:43:39 crc kubenswrapper[4857]: I1201 22:43:39.482859 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/pull/0.log" Dec 01 22:43:39 crc kubenswrapper[4857]: I1201 22:43:39.503514 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/pull/0.log" Dec 01 22:43:39 crc kubenswrapper[4857]: I1201 22:43:39.518764 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/util/0.log" Dec 01 22:43:39 crc kubenswrapper[4857]: I1201 22:43:39.690315 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/util/0.log" Dec 01 22:43:39 crc kubenswrapper[4857]: I1201 22:43:39.704922 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/pull/0.log" Dec 01 22:43:39 crc kubenswrapper[4857]: I1201 22:43:39.717139 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/extract/0.log" Dec 01 22:43:39 crc kubenswrapper[4857]: I1201 22:43:39.882421 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-rqfzk_80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4/kube-rbac-proxy/0.log" Dec 01 22:43:39 crc kubenswrapper[4857]: I1201 22:43:39.938496 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-rqfzk_80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4/manager/0.log" Dec 01 22:43:40 crc kubenswrapper[4857]: I1201 22:43:40.012025 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-5gqcm_d3a93b19-a249-499a-89fc-9b2a1a69c1e0/kube-rbac-proxy/0.log" Dec 01 22:43:40 crc kubenswrapper[4857]: I1201 22:43:40.176620 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-5gqcm_d3a93b19-a249-499a-89fc-9b2a1a69c1e0/manager/0.log" Dec 01 22:43:40 crc kubenswrapper[4857]: I1201 22:43:40.186806 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-dz5k5_393fd1a5-e3b9-44d6-a9ef-1447a0f9f465/kube-rbac-proxy/0.log" Dec 01 22:43:40 crc kubenswrapper[4857]: I1201 22:43:40.234506 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-dz5k5_393fd1a5-e3b9-44d6-a9ef-1447a0f9f465/manager/0.log" Dec 01 22:43:40 crc kubenswrapper[4857]: I1201 22:43:40.386120 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-9qwbq_df91f733-711d-43b0-98a0-b3fa6b119fbe/kube-rbac-proxy/0.log" Dec 01 22:43:40 crc kubenswrapper[4857]: I1201 22:43:40.417496 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-9qwbq_df91f733-711d-43b0-98a0-b3fa6b119fbe/manager/0.log" Dec 01 22:43:40 crc kubenswrapper[4857]: I1201 22:43:40.591709 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-zkqw9_068a21df-001f-45e6-9946-68f67bad7d43/kube-rbac-proxy/0.log" Dec 01 22:43:40 crc kubenswrapper[4857]: I1201 22:43:40.645291 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-zkqw9_068a21df-001f-45e6-9946-68f67bad7d43/manager/0.log" Dec 01 22:43:40 crc kubenswrapper[4857]: I1201 22:43:40.711917 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-4g68b_b972200d-235d-4ae5-a71d-dd206e4f6025/kube-rbac-proxy/0.log" Dec 01 22:43:40 crc kubenswrapper[4857]: I1201 22:43:40.787000 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-4g68b_b972200d-235d-4ae5-a71d-dd206e4f6025/manager/0.log" Dec 01 22:43:40 crc kubenswrapper[4857]: I1201 22:43:40.888579 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-54xtc_23e7348c-65d9-4c4b-9c1e-d86cc3b55bff/kube-rbac-proxy/0.log" Dec 01 22:43:41 crc kubenswrapper[4857]: I1201 22:43:41.051183 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-j277f_5dd6bfb5-2edc-49b4-8877-85ab5722cdf0/kube-rbac-proxy/0.log" Dec 01 22:43:41 crc kubenswrapper[4857]: I1201 22:43:41.068573 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-54xtc_23e7348c-65d9-4c4b-9c1e-d86cc3b55bff/manager/0.log" Dec 01 22:43:41 crc kubenswrapper[4857]: I1201 22:43:41.103561 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-j277f_5dd6bfb5-2edc-49b4-8877-85ab5722cdf0/manager/0.log" Dec 01 22:43:41 crc kubenswrapper[4857]: I1201 22:43:41.258031 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-vtcx7_dcef9388-5d9b-4172-bb80-ec69b3e98dd0/kube-rbac-proxy/0.log" Dec 01 22:43:41 crc kubenswrapper[4857]: I1201 22:43:41.311474 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-vtcx7_dcef9388-5d9b-4172-bb80-ec69b3e98dd0/manager/0.log" Dec 01 22:43:41 crc kubenswrapper[4857]: I1201 22:43:41.421071 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-b7ksb_92c43db5-8128-4269-9ad4-a263de88f450/kube-rbac-proxy/0.log" Dec 01 22:43:41 crc kubenswrapper[4857]: I1201 22:43:41.474771 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-b7ksb_92c43db5-8128-4269-9ad4-a263de88f450/manager/0.log" Dec 01 22:43:41 crc kubenswrapper[4857]: I1201 22:43:41.546399 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-f9d2t_b13a13cc-0689-480a-bc95-e19aba327f9c/kube-rbac-proxy/0.log" Dec 01 22:43:41 crc kubenswrapper[4857]: I1201 22:43:41.643307 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-f9d2t_b13a13cc-0689-480a-bc95-e19aba327f9c/manager/0.log" Dec 01 22:43:41 crc kubenswrapper[4857]: I1201 22:43:41.716887 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-p6v4v_188a7d67-ffec-4608-9b5f-cc33972e6fed/kube-rbac-proxy/0.log" Dec 01 22:43:41 crc kubenswrapper[4857]: I1201 22:43:41.880816 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-p6v4v_188a7d67-ffec-4608-9b5f-cc33972e6fed/manager/0.log" Dec 01 22:43:41 crc kubenswrapper[4857]: I1201 22:43:41.974353 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vbvnt_de53c246-6e20-4b67-be68-1c5356fd46d4/kube-rbac-proxy/0.log" Dec 01 22:43:42 crc kubenswrapper[4857]: I1201 22:43:42.064861 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vbvnt_de53c246-6e20-4b67-be68-1c5356fd46d4/manager/0.log" Dec 01 22:43:42 crc kubenswrapper[4857]: I1201 22:43:42.233240 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-cncd6_246a4bf0-515b-46f8-bd7f-89e7e961d774/kube-rbac-proxy/0.log" Dec 01 22:43:42 crc kubenswrapper[4857]: I1201 22:43:42.391150 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-cncd6_246a4bf0-515b-46f8-bd7f-89e7e961d774/manager/0.log" Dec 01 22:43:42 crc kubenswrapper[4857]: I1201 22:43:42.532675 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c_1c2d54cf-e53d-483d-9b00-6f03858806ab/manager/0.log" Dec 01 22:43:42 crc kubenswrapper[4857]: I1201 22:43:42.534426 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c_1c2d54cf-e53d-483d-9b00-6f03858806ab/kube-rbac-proxy/0.log" Dec 01 22:43:42 crc kubenswrapper[4857]: I1201 22:43:42.869441 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-m6lxm_f8b95808-7d34-4097-bac8-f7e081d9c039/registry-server/0.log" Dec 01 22:43:42 crc kubenswrapper[4857]: I1201 22:43:42.915582 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-9f7fd9cfd-66kxd_1e12f76f-2a0a-46d8-aec6-6d9a7cf7e0dc/operator/0.log" Dec 01 22:43:43 crc kubenswrapper[4857]: I1201 22:43:43.074798 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-7vmng_f6e43ef6-1269-40f5-af4d-fe2f30dea319/kube-rbac-proxy/0.log" Dec 01 22:43:43 crc kubenswrapper[4857]: I1201 22:43:43.138528 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-7vmng_f6e43ef6-1269-40f5-af4d-fe2f30dea319/manager/0.log" Dec 01 22:43:43 crc kubenswrapper[4857]: I1201 22:43:43.240311 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-mww86_c5ab847c-10d3-482d-8c70-e3ab439e7b5d/kube-rbac-proxy/0.log" Dec 01 22:43:43 crc kubenswrapper[4857]: I1201 22:43:43.267414 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-mww86_c5ab847c-10d3-482d-8c70-e3ab439e7b5d/manager/0.log" Dec 01 22:43:43 crc kubenswrapper[4857]: I1201 22:43:43.505669 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-wjnzv_cdd278ff-9eb6-45f7-977d-02944f6f312a/kube-rbac-proxy/0.log" Dec 01 22:43:43 crc kubenswrapper[4857]: I1201 22:43:43.539735 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-gdqjh_1d490d1b-558b-49fb-9b0f-65d0f82b0148/operator/0.log" Dec 01 22:43:43 crc kubenswrapper[4857]: I1201 22:43:43.581849 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-wjnzv_cdd278ff-9eb6-45f7-977d-02944f6f312a/manager/0.log" Dec 01 22:43:43 crc kubenswrapper[4857]: I1201 22:43:43.717885 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f765fc7d6-x797d_ff230e58-1711-4638-89d6-daf91fb3bfc1/kube-rbac-proxy/0.log" Dec 01 22:43:43 crc kubenswrapper[4857]: I1201 22:43:43.942261 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f765fc7d6-x797d_ff230e58-1711-4638-89d6-daf91fb3bfc1/manager/0.log" Dec 01 22:43:43 crc kubenswrapper[4857]: I1201 22:43:43.942604 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-777b97d487-s7g6x_0a4c6432-bb6c-41b6-9981-ef6ac1354ae3/manager/0.log" Dec 01 22:43:43 crc kubenswrapper[4857]: I1201 22:43:43.983986 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wccvn_4c8d087e-e00c-4dad-87c7-06ab1017abaf/kube-rbac-proxy/0.log" Dec 01 22:43:43 crc kubenswrapper[4857]: I1201 22:43:43.989447 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wccvn_4c8d087e-e00c-4dad-87c7-06ab1017abaf/manager/0.log" Dec 01 22:43:44 crc kubenswrapper[4857]: I1201 22:43:44.135904 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-vh887_8a6276de-e994-4f62-a625-906c9a5f4f5d/kube-rbac-proxy/0.log" Dec 01 22:43:44 crc kubenswrapper[4857]: I1201 22:43:44.149820 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-vh887_8a6276de-e994-4f62-a625-906c9a5f4f5d/manager/0.log" Dec 01 22:43:48 crc kubenswrapper[4857]: I1201 22:43:48.834820 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:43:48 crc kubenswrapper[4857]: E1201 22:43:48.836023 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:44:02 crc kubenswrapper[4857]: I1201 22:44:02.834779 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:44:03 crc kubenswrapper[4857]: I1201 22:44:03.287528 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"4d908912b33495ea153b8b1fbe5c79474599457200a6669673854ac4fb373bcb"} Dec 01 22:44:04 crc kubenswrapper[4857]: I1201 22:44:04.531768 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-8mmlp_1cb66d76-6c0a-48c9-a0a8-54a60f7aae01/control-plane-machine-set-operator/0.log" Dec 01 22:44:04 crc kubenswrapper[4857]: I1201 22:44:04.746268 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-2bwxl_00d4b52b-5372-48a1-a1b4-6c17e18cc7a4/machine-api-operator/0.log" Dec 01 22:44:04 crc kubenswrapper[4857]: I1201 22:44:04.750927 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-2bwxl_00d4b52b-5372-48a1-a1b4-6c17e18cc7a4/kube-rbac-proxy/0.log" Dec 01 22:44:19 crc kubenswrapper[4857]: I1201 22:44:19.216378 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-wnq4k_e8804d0a-50d0-4b87-bc77-a8c8229ccf1d/cert-manager-controller/0.log" Dec 01 22:44:19 crc kubenswrapper[4857]: I1201 22:44:19.332827 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-lcww8_84a659be-d476-4ba1-820e-c5279704aec5/cert-manager-cainjector/0.log" Dec 01 22:44:19 crc kubenswrapper[4857]: I1201 22:44:19.400390 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-5gv57_64fd8572-2a45-422e-a413-49314dc52a55/cert-manager-webhook/0.log" Dec 01 22:44:32 crc kubenswrapper[4857]: I1201 22:44:32.971902 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-78hdv_0764bcb8-a3f2-46cd-bceb-f08b51012b7d/nmstate-console-plugin/0.log" Dec 01 22:44:33 crc kubenswrapper[4857]: I1201 22:44:33.152203 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-grx52_e37b046b-176a-4bc3-bcbc-3093fe08b256/nmstate-handler/0.log" Dec 01 22:44:33 crc kubenswrapper[4857]: I1201 22:44:33.161134 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-m49jm_c7894b0e-91be-4ed2-810b-f9e24df87fae/kube-rbac-proxy/0.log" Dec 01 22:44:33 crc kubenswrapper[4857]: I1201 22:44:33.243494 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-m49jm_c7894b0e-91be-4ed2-810b-f9e24df87fae/nmstate-metrics/0.log" Dec 01 22:44:33 crc kubenswrapper[4857]: I1201 22:44:33.348605 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-m8q4r_8c989c5f-b814-437a-bdca-86e63044ce91/nmstate-operator/0.log" Dec 01 22:44:33 crc kubenswrapper[4857]: I1201 22:44:33.425465 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-bdx6w_79431306-c7c5-4c07-8b52-ec94fe45e6da/nmstate-webhook/0.log" Dec 01 22:44:51 crc kubenswrapper[4857]: I1201 22:44:51.194224 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-2gzlv_28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4/kube-rbac-proxy/0.log" Dec 01 22:44:51 crc kubenswrapper[4857]: I1201 22:44:51.236507 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-2gzlv_28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4/controller/0.log" Dec 01 22:44:51 crc kubenswrapper[4857]: I1201 22:44:51.374145 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-x6nqn_ee319d44-c1a3-4325-835f-3d9fc6a187fd/frr-k8s-webhook-server/0.log" Dec 01 22:44:51 crc kubenswrapper[4857]: I1201 22:44:51.442124 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-frr-files/0.log" Dec 01 22:44:51 crc kubenswrapper[4857]: I1201 22:44:51.615641 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-reloader/0.log" Dec 01 22:44:51 crc kubenswrapper[4857]: I1201 22:44:51.623931 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-frr-files/0.log" Dec 01 22:44:51 crc kubenswrapper[4857]: I1201 22:44:51.632006 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-metrics/0.log" Dec 01 22:44:51 crc kubenswrapper[4857]: I1201 22:44:51.656023 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-reloader/0.log" Dec 01 22:44:51 crc kubenswrapper[4857]: I1201 22:44:51.854800 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-frr-files/0.log" Dec 01 22:44:51 crc kubenswrapper[4857]: I1201 22:44:51.862383 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-reloader/0.log" Dec 01 22:44:51 crc kubenswrapper[4857]: I1201 22:44:51.883121 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-metrics/0.log" Dec 01 22:44:51 crc kubenswrapper[4857]: I1201 22:44:51.887088 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-metrics/0.log" Dec 01 22:44:52 crc kubenswrapper[4857]: I1201 22:44:52.748984 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-frr-files/0.log" Dec 01 22:44:52 crc kubenswrapper[4857]: I1201 22:44:52.773357 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-metrics/0.log" Dec 01 22:44:52 crc kubenswrapper[4857]: I1201 22:44:52.800998 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-reloader/0.log" Dec 01 22:44:52 crc kubenswrapper[4857]: I1201 22:44:52.812103 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/controller/0.log" Dec 01 22:44:52 crc kubenswrapper[4857]: I1201 22:44:52.952762 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/frr-metrics/0.log" Dec 01 22:44:53 crc kubenswrapper[4857]: I1201 22:44:53.005061 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/kube-rbac-proxy/0.log" Dec 01 22:44:53 crc kubenswrapper[4857]: I1201 22:44:53.034882 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/kube-rbac-proxy-frr/0.log" Dec 01 22:44:53 crc kubenswrapper[4857]: I1201 22:44:53.205700 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/reloader/0.log" Dec 01 22:44:53 crc kubenswrapper[4857]: I1201 22:44:53.227568 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-75c9cd8c7-tt4cx_8eee3586-4292-4165-81c4-496c2ee8ab9c/manager/0.log" Dec 01 22:44:53 crc kubenswrapper[4857]: I1201 22:44:53.388683 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-75b5f488d8-k8rxs_bd5afa07-4d43-4dce-8292-58ad63e6362a/webhook-server/0.log" Dec 01 22:44:53 crc kubenswrapper[4857]: I1201 22:44:53.617217 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r64qq_c6620f92-fe9f-4dde-a758-b46475bd8db8/kube-rbac-proxy/0.log" Dec 01 22:44:54 crc kubenswrapper[4857]: I1201 22:44:54.077748 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r64qq_c6620f92-fe9f-4dde-a758-b46475bd8db8/speaker/0.log" Dec 01 22:44:54 crc kubenswrapper[4857]: I1201 22:44:54.426032 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/frr/0.log" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.209327 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs"] Dec 01 22:45:00 crc kubenswrapper[4857]: E1201 22:45:00.210534 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9a5cc2c-8b4f-412b-8eaa-90084d966507" containerName="container-00" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.210565 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9a5cc2c-8b4f-412b-8eaa-90084d966507" containerName="container-00" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.210903 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9a5cc2c-8b4f-412b-8eaa-90084d966507" containerName="container-00" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.211863 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.214831 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.217484 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.221318 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs"] Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.303163 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9mn9\" (UniqueName: \"kubernetes.io/projected/3d0043b7-3b3f-4f9e-93cc-94053841f04e-kube-api-access-j9mn9\") pod \"collect-profiles-29410485-7bhcs\" (UID: \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.303456 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d0043b7-3b3f-4f9e-93cc-94053841f04e-config-volume\") pod \"collect-profiles-29410485-7bhcs\" (UID: \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.303485 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d0043b7-3b3f-4f9e-93cc-94053841f04e-secret-volume\") pod \"collect-profiles-29410485-7bhcs\" (UID: \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.404864 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d0043b7-3b3f-4f9e-93cc-94053841f04e-config-volume\") pod \"collect-profiles-29410485-7bhcs\" (UID: \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.404927 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d0043b7-3b3f-4f9e-93cc-94053841f04e-secret-volume\") pod \"collect-profiles-29410485-7bhcs\" (UID: \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.405116 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9mn9\" (UniqueName: \"kubernetes.io/projected/3d0043b7-3b3f-4f9e-93cc-94053841f04e-kube-api-access-j9mn9\") pod \"collect-profiles-29410485-7bhcs\" (UID: \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.407310 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d0043b7-3b3f-4f9e-93cc-94053841f04e-config-volume\") pod \"collect-profiles-29410485-7bhcs\" (UID: \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.418860 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d0043b7-3b3f-4f9e-93cc-94053841f04e-secret-volume\") pod \"collect-profiles-29410485-7bhcs\" (UID: \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.430035 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9mn9\" (UniqueName: \"kubernetes.io/projected/3d0043b7-3b3f-4f9e-93cc-94053841f04e-kube-api-access-j9mn9\") pod \"collect-profiles-29410485-7bhcs\" (UID: \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:00 crc kubenswrapper[4857]: I1201 22:45:00.534069 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:01 crc kubenswrapper[4857]: I1201 22:45:01.074245 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs"] Dec 01 22:45:01 crc kubenswrapper[4857]: W1201 22:45:01.077011 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d0043b7_3b3f_4f9e_93cc_94053841f04e.slice/crio-4b64e1ec9c97e7a61ddeb36f34941f9a88810ed1afcbf9668d560c16dd87ea3d WatchSource:0}: Error finding container 4b64e1ec9c97e7a61ddeb36f34941f9a88810ed1afcbf9668d560c16dd87ea3d: Status 404 returned error can't find the container with id 4b64e1ec9c97e7a61ddeb36f34941f9a88810ed1afcbf9668d560c16dd87ea3d Dec 01 22:45:01 crc kubenswrapper[4857]: I1201 22:45:01.963096 4857 generic.go:334] "Generic (PLEG): container finished" podID="3d0043b7-3b3f-4f9e-93cc-94053841f04e" containerID="a029de25f9fbf4b0218920b074e06d92685f322218639b04da040865c2a154b4" exitCode=0 Dec 01 22:45:01 crc kubenswrapper[4857]: I1201 22:45:01.963315 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" event={"ID":"3d0043b7-3b3f-4f9e-93cc-94053841f04e","Type":"ContainerDied","Data":"a029de25f9fbf4b0218920b074e06d92685f322218639b04da040865c2a154b4"} Dec 01 22:45:01 crc kubenswrapper[4857]: I1201 22:45:01.963738 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" event={"ID":"3d0043b7-3b3f-4f9e-93cc-94053841f04e","Type":"ContainerStarted","Data":"4b64e1ec9c97e7a61ddeb36f34941f9a88810ed1afcbf9668d560c16dd87ea3d"} Dec 01 22:45:03 crc kubenswrapper[4857]: I1201 22:45:03.995154 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" event={"ID":"3d0043b7-3b3f-4f9e-93cc-94053841f04e","Type":"ContainerDied","Data":"4b64e1ec9c97e7a61ddeb36f34941f9a88810ed1afcbf9668d560c16dd87ea3d"} Dec 01 22:45:03 crc kubenswrapper[4857]: I1201 22:45:03.995524 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b64e1ec9c97e7a61ddeb36f34941f9a88810ed1afcbf9668d560c16dd87ea3d" Dec 01 22:45:04 crc kubenswrapper[4857]: I1201 22:45:04.111651 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:04 crc kubenswrapper[4857]: I1201 22:45:04.186782 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d0043b7-3b3f-4f9e-93cc-94053841f04e-secret-volume\") pod \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\" (UID: \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\") " Dec 01 22:45:04 crc kubenswrapper[4857]: I1201 22:45:04.187471 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9mn9\" (UniqueName: \"kubernetes.io/projected/3d0043b7-3b3f-4f9e-93cc-94053841f04e-kube-api-access-j9mn9\") pod \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\" (UID: \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\") " Dec 01 22:45:04 crc kubenswrapper[4857]: I1201 22:45:04.187748 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d0043b7-3b3f-4f9e-93cc-94053841f04e-config-volume\") pod \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\" (UID: \"3d0043b7-3b3f-4f9e-93cc-94053841f04e\") " Dec 01 22:45:04 crc kubenswrapper[4857]: I1201 22:45:04.188783 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d0043b7-3b3f-4f9e-93cc-94053841f04e-config-volume" (OuterVolumeSpecName: "config-volume") pod "3d0043b7-3b3f-4f9e-93cc-94053841f04e" (UID: "3d0043b7-3b3f-4f9e-93cc-94053841f04e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 22:45:04 crc kubenswrapper[4857]: I1201 22:45:04.197197 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d0043b7-3b3f-4f9e-93cc-94053841f04e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3d0043b7-3b3f-4f9e-93cc-94053841f04e" (UID: "3d0043b7-3b3f-4f9e-93cc-94053841f04e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 22:45:04 crc kubenswrapper[4857]: I1201 22:45:04.197656 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d0043b7-3b3f-4f9e-93cc-94053841f04e-kube-api-access-j9mn9" (OuterVolumeSpecName: "kube-api-access-j9mn9") pod "3d0043b7-3b3f-4f9e-93cc-94053841f04e" (UID: "3d0043b7-3b3f-4f9e-93cc-94053841f04e"). InnerVolumeSpecName "kube-api-access-j9mn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:45:04 crc kubenswrapper[4857]: I1201 22:45:04.290712 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9mn9\" (UniqueName: \"kubernetes.io/projected/3d0043b7-3b3f-4f9e-93cc-94053841f04e-kube-api-access-j9mn9\") on node \"crc\" DevicePath \"\"" Dec 01 22:45:04 crc kubenswrapper[4857]: I1201 22:45:04.290772 4857 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d0043b7-3b3f-4f9e-93cc-94053841f04e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 22:45:04 crc kubenswrapper[4857]: I1201 22:45:04.290789 4857 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d0043b7-3b3f-4f9e-93cc-94053841f04e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 22:45:05 crc kubenswrapper[4857]: I1201 22:45:05.005259 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410485-7bhcs" Dec 01 22:45:05 crc kubenswrapper[4857]: I1201 22:45:05.242704 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2"] Dec 01 22:45:05 crc kubenswrapper[4857]: I1201 22:45:05.273639 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410440-nl2z2"] Dec 01 22:45:05 crc kubenswrapper[4857]: I1201 22:45:05.847347 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1112bd6e-216a-48cb-9505-b1caf7d178cd" path="/var/lib/kubelet/pods/1112bd6e-216a-48cb-9505-b1caf7d178cd/volumes" Dec 01 22:45:09 crc kubenswrapper[4857]: I1201 22:45:09.181899 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/util/0.log" Dec 01 22:45:09 crc kubenswrapper[4857]: I1201 22:45:09.376897 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/util/0.log" Dec 01 22:45:09 crc kubenswrapper[4857]: I1201 22:45:09.377161 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/pull/0.log" Dec 01 22:45:09 crc kubenswrapper[4857]: I1201 22:45:09.398633 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/pull/0.log" Dec 01 22:45:09 crc kubenswrapper[4857]: I1201 22:45:09.583899 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/pull/0.log" Dec 01 22:45:09 crc kubenswrapper[4857]: I1201 22:45:09.596105 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/util/0.log" Dec 01 22:45:09 crc kubenswrapper[4857]: I1201 22:45:09.620068 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/extract/0.log" Dec 01 22:45:09 crc kubenswrapper[4857]: I1201 22:45:09.809357 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/util/0.log" Dec 01 22:45:09 crc kubenswrapper[4857]: I1201 22:45:09.941076 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/pull/0.log" Dec 01 22:45:09 crc kubenswrapper[4857]: I1201 22:45:09.952512 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/pull/0.log" Dec 01 22:45:09 crc kubenswrapper[4857]: I1201 22:45:09.978031 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/util/0.log" Dec 01 22:45:10 crc kubenswrapper[4857]: I1201 22:45:10.132122 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/extract/0.log" Dec 01 22:45:10 crc kubenswrapper[4857]: I1201 22:45:10.138973 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/util/0.log" Dec 01 22:45:10 crc kubenswrapper[4857]: I1201 22:45:10.174265 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/pull/0.log" Dec 01 22:45:10 crc kubenswrapper[4857]: I1201 22:45:10.340956 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/util/0.log" Dec 01 22:45:10 crc kubenswrapper[4857]: I1201 22:45:10.521806 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/pull/0.log" Dec 01 22:45:10 crc kubenswrapper[4857]: I1201 22:45:10.539417 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/pull/0.log" Dec 01 22:45:10 crc kubenswrapper[4857]: I1201 22:45:10.569189 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/util/0.log" Dec 01 22:45:10 crc kubenswrapper[4857]: I1201 22:45:10.733153 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/pull/0.log" Dec 01 22:45:10 crc kubenswrapper[4857]: I1201 22:45:10.738917 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/util/0.log" Dec 01 22:45:10 crc kubenswrapper[4857]: I1201 22:45:10.742297 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/extract/0.log" Dec 01 22:45:10 crc kubenswrapper[4857]: I1201 22:45:10.901493 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/extract-utilities/0.log" Dec 01 22:45:11 crc kubenswrapper[4857]: I1201 22:45:11.040447 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/extract-content/0.log" Dec 01 22:45:11 crc kubenswrapper[4857]: I1201 22:45:11.083538 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/extract-content/0.log" Dec 01 22:45:11 crc kubenswrapper[4857]: I1201 22:45:11.098208 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/extract-utilities/0.log" Dec 01 22:45:11 crc kubenswrapper[4857]: I1201 22:45:11.271654 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/extract-content/0.log" Dec 01 22:45:11 crc kubenswrapper[4857]: I1201 22:45:11.273596 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/extract-utilities/0.log" Dec 01 22:45:11 crc kubenswrapper[4857]: I1201 22:45:11.480788 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/extract-utilities/0.log" Dec 01 22:45:11 crc kubenswrapper[4857]: I1201 22:45:11.651808 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/extract-utilities/0.log" Dec 01 22:45:11 crc kubenswrapper[4857]: I1201 22:45:11.682901 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/registry-server/0.log" Dec 01 22:45:11 crc kubenswrapper[4857]: I1201 22:45:11.767644 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/extract-content/0.log" Dec 01 22:45:11 crc kubenswrapper[4857]: I1201 22:45:11.785605 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/extract-content/0.log" Dec 01 22:45:11 crc kubenswrapper[4857]: I1201 22:45:11.911693 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/extract-content/0.log" Dec 01 22:45:11 crc kubenswrapper[4857]: I1201 22:45:11.926363 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/extract-utilities/0.log" Dec 01 22:45:12 crc kubenswrapper[4857]: I1201 22:45:12.109732 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-d5gjm_9fe4a11e-f252-4f61-a8f4-d513b2b40643/marketplace-operator/0.log" Dec 01 22:45:12 crc kubenswrapper[4857]: I1201 22:45:12.237369 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/extract-utilities/0.log" Dec 01 22:45:12 crc kubenswrapper[4857]: I1201 22:45:12.489526 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/registry-server/0.log" Dec 01 22:45:12 crc kubenswrapper[4857]: I1201 22:45:12.494327 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/extract-utilities/0.log" Dec 01 22:45:12 crc kubenswrapper[4857]: I1201 22:45:12.513366 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/extract-content/0.log" Dec 01 22:45:12 crc kubenswrapper[4857]: I1201 22:45:12.522797 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/extract-content/0.log" Dec 01 22:45:12 crc kubenswrapper[4857]: I1201 22:45:12.846401 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/extract-content/0.log" Dec 01 22:45:13 crc kubenswrapper[4857]: I1201 22:45:13.013779 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/extract-utilities/0.log" Dec 01 22:45:13 crc kubenswrapper[4857]: I1201 22:45:13.039352 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/extract-utilities/0.log" Dec 01 22:45:13 crc kubenswrapper[4857]: I1201 22:45:13.054404 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/registry-server/0.log" Dec 01 22:45:13 crc kubenswrapper[4857]: I1201 22:45:13.182549 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/extract-utilities/0.log" Dec 01 22:45:13 crc kubenswrapper[4857]: I1201 22:45:13.194945 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/extract-content/0.log" Dec 01 22:45:13 crc kubenswrapper[4857]: I1201 22:45:13.200565 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/extract-content/0.log" Dec 01 22:45:13 crc kubenswrapper[4857]: I1201 22:45:13.382626 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/extract-utilities/0.log" Dec 01 22:45:13 crc kubenswrapper[4857]: I1201 22:45:13.386694 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/extract-content/0.log" Dec 01 22:45:13 crc kubenswrapper[4857]: I1201 22:45:13.896172 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/registry-server/0.log" Dec 01 22:45:25 crc kubenswrapper[4857]: I1201 22:45:25.064443 4857 scope.go:117] "RemoveContainer" containerID="d964aba06e670c29e4ba369c125f0d4a923af04e21298593213068e8a85a1bef" Dec 01 22:45:30 crc kubenswrapper[4857]: I1201 22:45:30.670855 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-tqws8_45d679d3-a9f8-402a-a8cf-4f9913ea40a1/prometheus-operator/0.log" Dec 01 22:45:30 crc kubenswrapper[4857]: I1201 22:45:30.934355 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-744b77886-6c586_203588e4-df3f-407e-9454-9cff35f281e5/prometheus-operator-admission-webhook/0.log" Dec 01 22:45:30 crc kubenswrapper[4857]: I1201 22:45:30.998630 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-744b77886-qczzg_e212d47c-24b3-44cb-a1de-3a291777358d/prometheus-operator-admission-webhook/0.log" Dec 01 22:45:31 crc kubenswrapper[4857]: I1201 22:45:31.110481 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-bxmst_f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f/operator/0.log" Dec 01 22:45:31 crc kubenswrapper[4857]: I1201 22:45:31.199549 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-26djf_f5a63bac-285c-401a-a975-3079cbad38cf/perses-operator/0.log" Dec 01 22:46:27 crc kubenswrapper[4857]: I1201 22:46:27.806708 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:46:27 crc kubenswrapper[4857]: I1201 22:46:27.807199 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:46:57 crc kubenswrapper[4857]: I1201 22:46:57.807062 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:46:57 crc kubenswrapper[4857]: I1201 22:46:57.807591 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:47:11 crc kubenswrapper[4857]: I1201 22:47:11.019210 4857 generic.go:334] "Generic (PLEG): container finished" podID="c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d" containerID="27e63ca2383fb545df3332964fe2e89e05b70c3cfa909a0777cc96913abc4fa8" exitCode=0 Dec 01 22:47:11 crc kubenswrapper[4857]: I1201 22:47:11.019275 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgsk4/must-gather-glvmw" event={"ID":"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d","Type":"ContainerDied","Data":"27e63ca2383fb545df3332964fe2e89e05b70c3cfa909a0777cc96913abc4fa8"} Dec 01 22:47:11 crc kubenswrapper[4857]: I1201 22:47:11.022251 4857 scope.go:117] "RemoveContainer" containerID="27e63ca2383fb545df3332964fe2e89e05b70c3cfa909a0777cc96913abc4fa8" Dec 01 22:47:11 crc kubenswrapper[4857]: I1201 22:47:11.740750 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mgsk4_must-gather-glvmw_c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d/gather/0.log" Dec 01 22:47:19 crc kubenswrapper[4857]: I1201 22:47:19.937596 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mgsk4/must-gather-glvmw"] Dec 01 22:47:19 crc kubenswrapper[4857]: I1201 22:47:19.938599 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mgsk4/must-gather-glvmw" podUID="c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d" containerName="copy" containerID="cri-o://8371467cd8fd4561b871b544cd3c878f9dfc722c5ecfab0507e19e49ecc732e7" gracePeriod=2 Dec 01 22:47:19 crc kubenswrapper[4857]: I1201 22:47:19.952352 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mgsk4/must-gather-glvmw"] Dec 01 22:47:20 crc kubenswrapper[4857]: I1201 22:47:20.260196 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mgsk4_must-gather-glvmw_c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d/copy/0.log" Dec 01 22:47:20 crc kubenswrapper[4857]: I1201 22:47:20.262297 4857 generic.go:334] "Generic (PLEG): container finished" podID="c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d" containerID="8371467cd8fd4561b871b544cd3c878f9dfc722c5ecfab0507e19e49ecc732e7" exitCode=143 Dec 01 22:47:20 crc kubenswrapper[4857]: I1201 22:47:20.507191 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mgsk4_must-gather-glvmw_c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d/copy/0.log" Dec 01 22:47:20 crc kubenswrapper[4857]: I1201 22:47:20.507905 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgsk4/must-gather-glvmw" Dec 01 22:47:20 crc kubenswrapper[4857]: I1201 22:47:20.555482 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jscz\" (UniqueName: \"kubernetes.io/projected/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d-kube-api-access-9jscz\") pod \"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d\" (UID: \"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d\") " Dec 01 22:47:20 crc kubenswrapper[4857]: I1201 22:47:20.564413 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d-kube-api-access-9jscz" (OuterVolumeSpecName: "kube-api-access-9jscz") pod "c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d" (UID: "c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d"). InnerVolumeSpecName "kube-api-access-9jscz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:47:20 crc kubenswrapper[4857]: I1201 22:47:20.658087 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d-must-gather-output\") pod \"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d\" (UID: \"c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d\") " Dec 01 22:47:20 crc kubenswrapper[4857]: I1201 22:47:20.658775 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jscz\" (UniqueName: \"kubernetes.io/projected/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d-kube-api-access-9jscz\") on node \"crc\" DevicePath \"\"" Dec 01 22:47:20 crc kubenswrapper[4857]: I1201 22:47:20.833254 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d" (UID: "c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:47:20 crc kubenswrapper[4857]: I1201 22:47:20.863678 4857 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 22:47:21 crc kubenswrapper[4857]: I1201 22:47:21.274867 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mgsk4_must-gather-glvmw_c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d/copy/0.log" Dec 01 22:47:21 crc kubenswrapper[4857]: I1201 22:47:21.275718 4857 scope.go:117] "RemoveContainer" containerID="8371467cd8fd4561b871b544cd3c878f9dfc722c5ecfab0507e19e49ecc732e7" Dec 01 22:47:21 crc kubenswrapper[4857]: I1201 22:47:21.275834 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgsk4/must-gather-glvmw" Dec 01 22:47:21 crc kubenswrapper[4857]: I1201 22:47:21.304763 4857 scope.go:117] "RemoveContainer" containerID="27e63ca2383fb545df3332964fe2e89e05b70c3cfa909a0777cc96913abc4fa8" Dec 01 22:47:21 crc kubenswrapper[4857]: I1201 22:47:21.854001 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d" path="/var/lib/kubelet/pods/c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d/volumes" Dec 01 22:47:25 crc kubenswrapper[4857]: I1201 22:47:25.158725 4857 scope.go:117] "RemoveContainer" containerID="96dab71e7508cbbcc6b7a5e4b5435f8eb47dde28bd73498bec459040a718635b" Dec 01 22:47:25 crc kubenswrapper[4857]: I1201 22:47:25.199293 4857 scope.go:117] "RemoveContainer" containerID="8b2d40e9d8c740048854c7f8b945176e480276c175876206ecb78b9a568d3e73" Dec 01 22:47:25 crc kubenswrapper[4857]: I1201 22:47:25.275851 4857 scope.go:117] "RemoveContainer" containerID="a64e60719be23537d57fe5d08926f5a29e7df14f9c5142771be10997f18a2001" Dec 01 22:47:27 crc kubenswrapper[4857]: I1201 22:47:27.806370 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:47:27 crc kubenswrapper[4857]: I1201 22:47:27.806902 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:47:27 crc kubenswrapper[4857]: I1201 22:47:27.806995 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 22:47:27 crc kubenswrapper[4857]: I1201 22:47:27.808451 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4d908912b33495ea153b8b1fbe5c79474599457200a6669673854ac4fb373bcb"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 22:47:27 crc kubenswrapper[4857]: I1201 22:47:27.808604 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://4d908912b33495ea153b8b1fbe5c79474599457200a6669673854ac4fb373bcb" gracePeriod=600 Dec 01 22:47:28 crc kubenswrapper[4857]: I1201 22:47:28.407410 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="4d908912b33495ea153b8b1fbe5c79474599457200a6669673854ac4fb373bcb" exitCode=0 Dec 01 22:47:28 crc kubenswrapper[4857]: I1201 22:47:28.408129 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"4d908912b33495ea153b8b1fbe5c79474599457200a6669673854ac4fb373bcb"} Dec 01 22:47:28 crc kubenswrapper[4857]: I1201 22:47:28.408174 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4"} Dec 01 22:47:28 crc kubenswrapper[4857]: I1201 22:47:28.408198 4857 scope.go:117] "RemoveContainer" containerID="bedc6b29fe0f3450da0ec0c8f13dc3f54d3e5634707c188b70134a4998fc3104" Dec 01 22:48:25 crc kubenswrapper[4857]: I1201 22:48:25.385125 4857 scope.go:117] "RemoveContainer" containerID="d362ad4cec62e9907637ca86493a21bc47c59a19dda83e36c36a13efccb742ac" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.577061 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qn8cm"] Dec 01 22:49:56 crc kubenswrapper[4857]: E1201 22:49:56.578220 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d" containerName="gather" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.578246 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d" containerName="gather" Dec 01 22:49:56 crc kubenswrapper[4857]: E1201 22:49:56.578269 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d" containerName="copy" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.578278 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d" containerName="copy" Dec 01 22:49:56 crc kubenswrapper[4857]: E1201 22:49:56.578309 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d0043b7-3b3f-4f9e-93cc-94053841f04e" containerName="collect-profiles" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.578320 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d0043b7-3b3f-4f9e-93cc-94053841f04e" containerName="collect-profiles" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.578619 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d" containerName="gather" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.578640 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d0043b7-3b3f-4f9e-93cc-94053841f04e" containerName="collect-profiles" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.578660 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4f5ba1b-4f57-4afe-9d69-e7b59fe01a5d" containerName="copy" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.580635 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.595576 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qn8cm"] Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.668471 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3808ad8-b638-412a-b668-00d8e0e76528-utilities\") pod \"redhat-operators-qn8cm\" (UID: \"c3808ad8-b638-412a-b668-00d8e0e76528\") " pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.668578 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3808ad8-b638-412a-b668-00d8e0e76528-catalog-content\") pod \"redhat-operators-qn8cm\" (UID: \"c3808ad8-b638-412a-b668-00d8e0e76528\") " pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.668797 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97zf7\" (UniqueName: \"kubernetes.io/projected/c3808ad8-b638-412a-b668-00d8e0e76528-kube-api-access-97zf7\") pod \"redhat-operators-qn8cm\" (UID: \"c3808ad8-b638-412a-b668-00d8e0e76528\") " pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.770441 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97zf7\" (UniqueName: \"kubernetes.io/projected/c3808ad8-b638-412a-b668-00d8e0e76528-kube-api-access-97zf7\") pod \"redhat-operators-qn8cm\" (UID: \"c3808ad8-b638-412a-b668-00d8e0e76528\") " pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.770578 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3808ad8-b638-412a-b668-00d8e0e76528-utilities\") pod \"redhat-operators-qn8cm\" (UID: \"c3808ad8-b638-412a-b668-00d8e0e76528\") " pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.770622 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3808ad8-b638-412a-b668-00d8e0e76528-catalog-content\") pod \"redhat-operators-qn8cm\" (UID: \"c3808ad8-b638-412a-b668-00d8e0e76528\") " pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.771429 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3808ad8-b638-412a-b668-00d8e0e76528-utilities\") pod \"redhat-operators-qn8cm\" (UID: \"c3808ad8-b638-412a-b668-00d8e0e76528\") " pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.771438 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3808ad8-b638-412a-b668-00d8e0e76528-catalog-content\") pod \"redhat-operators-qn8cm\" (UID: \"c3808ad8-b638-412a-b668-00d8e0e76528\") " pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.799115 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97zf7\" (UniqueName: \"kubernetes.io/projected/c3808ad8-b638-412a-b668-00d8e0e76528-kube-api-access-97zf7\") pod \"redhat-operators-qn8cm\" (UID: \"c3808ad8-b638-412a-b668-00d8e0e76528\") " pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:49:56 crc kubenswrapper[4857]: I1201 22:49:56.917289 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:49:57 crc kubenswrapper[4857]: I1201 22:49:57.393424 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qn8cm"] Dec 01 22:49:57 crc kubenswrapper[4857]: I1201 22:49:57.566557 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn8cm" event={"ID":"c3808ad8-b638-412a-b668-00d8e0e76528","Type":"ContainerStarted","Data":"c325c3c2004ca0ad6999610dc1303f005eb1bd15eb42e619ca5f58afc03921f1"} Dec 01 22:49:57 crc kubenswrapper[4857]: I1201 22:49:57.805976 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:49:57 crc kubenswrapper[4857]: I1201 22:49:57.806064 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:49:58 crc kubenswrapper[4857]: I1201 22:49:58.588679 4857 generic.go:334] "Generic (PLEG): container finished" podID="c3808ad8-b638-412a-b668-00d8e0e76528" containerID="f3d3ac39225d449559d773a7399292029402199d14ef78eb458b64472f5d2178" exitCode=0 Dec 01 22:49:58 crc kubenswrapper[4857]: I1201 22:49:58.588834 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn8cm" event={"ID":"c3808ad8-b638-412a-b668-00d8e0e76528","Type":"ContainerDied","Data":"f3d3ac39225d449559d773a7399292029402199d14ef78eb458b64472f5d2178"} Dec 01 22:49:58 crc kubenswrapper[4857]: I1201 22:49:58.594945 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 22:50:00 crc kubenswrapper[4857]: I1201 22:50:00.628939 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn8cm" event={"ID":"c3808ad8-b638-412a-b668-00d8e0e76528","Type":"ContainerStarted","Data":"0b4cb7884b3af24aaaa58e7b361b641d080b5beb57b38287f795e6c3ce20cf69"} Dec 01 22:50:02 crc kubenswrapper[4857]: I1201 22:50:02.657977 4857 generic.go:334] "Generic (PLEG): container finished" podID="c3808ad8-b638-412a-b668-00d8e0e76528" containerID="0b4cb7884b3af24aaaa58e7b361b641d080b5beb57b38287f795e6c3ce20cf69" exitCode=0 Dec 01 22:50:02 crc kubenswrapper[4857]: I1201 22:50:02.658090 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn8cm" event={"ID":"c3808ad8-b638-412a-b668-00d8e0e76528","Type":"ContainerDied","Data":"0b4cb7884b3af24aaaa58e7b361b641d080b5beb57b38287f795e6c3ce20cf69"} Dec 01 22:50:04 crc kubenswrapper[4857]: I1201 22:50:04.683518 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn8cm" event={"ID":"c3808ad8-b638-412a-b668-00d8e0e76528","Type":"ContainerStarted","Data":"9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a"} Dec 01 22:50:04 crc kubenswrapper[4857]: I1201 22:50:04.711429 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qn8cm" podStartSLOduration=3.915237805 podStartE2EDuration="8.711389803s" podCreationTimestamp="2025-12-01 22:49:56 +0000 UTC" firstStartedPulling="2025-12-01 22:49:58.59418464 +0000 UTC m=+4517.084246997" lastFinishedPulling="2025-12-01 22:50:03.390336678 +0000 UTC m=+4521.880398995" observedRunningTime="2025-12-01 22:50:04.708552634 +0000 UTC m=+4523.198614971" watchObservedRunningTime="2025-12-01 22:50:04.711389803 +0000 UTC m=+4523.201452150" Dec 01 22:50:06 crc kubenswrapper[4857]: I1201 22:50:06.917652 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:50:06 crc kubenswrapper[4857]: I1201 22:50:06.918083 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:50:07 crc kubenswrapper[4857]: I1201 22:50:07.980195 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qn8cm" podUID="c3808ad8-b638-412a-b668-00d8e0e76528" containerName="registry-server" probeResult="failure" output=< Dec 01 22:50:07 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Dec 01 22:50:07 crc kubenswrapper[4857]: > Dec 01 22:50:16 crc kubenswrapper[4857]: I1201 22:50:16.984491 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:50:17 crc kubenswrapper[4857]: I1201 22:50:17.036537 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:50:17 crc kubenswrapper[4857]: I1201 22:50:17.246867 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qn8cm"] Dec 01 22:50:18 crc kubenswrapper[4857]: I1201 22:50:18.895650 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qn8cm" podUID="c3808ad8-b638-412a-b668-00d8e0e76528" containerName="registry-server" containerID="cri-o://9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a" gracePeriod=2 Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.510271 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.670878 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3808ad8-b638-412a-b668-00d8e0e76528-utilities\") pod \"c3808ad8-b638-412a-b668-00d8e0e76528\" (UID: \"c3808ad8-b638-412a-b668-00d8e0e76528\") " Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.671021 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97zf7\" (UniqueName: \"kubernetes.io/projected/c3808ad8-b638-412a-b668-00d8e0e76528-kube-api-access-97zf7\") pod \"c3808ad8-b638-412a-b668-00d8e0e76528\" (UID: \"c3808ad8-b638-412a-b668-00d8e0e76528\") " Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.671213 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3808ad8-b638-412a-b668-00d8e0e76528-catalog-content\") pod \"c3808ad8-b638-412a-b668-00d8e0e76528\" (UID: \"c3808ad8-b638-412a-b668-00d8e0e76528\") " Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.671836 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3808ad8-b638-412a-b668-00d8e0e76528-utilities" (OuterVolumeSpecName: "utilities") pod "c3808ad8-b638-412a-b668-00d8e0e76528" (UID: "c3808ad8-b638-412a-b668-00d8e0e76528"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.676848 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3808ad8-b638-412a-b668-00d8e0e76528-kube-api-access-97zf7" (OuterVolumeSpecName: "kube-api-access-97zf7") pod "c3808ad8-b638-412a-b668-00d8e0e76528" (UID: "c3808ad8-b638-412a-b668-00d8e0e76528"). InnerVolumeSpecName "kube-api-access-97zf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.689990 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3808ad8-b638-412a-b668-00d8e0e76528-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.690033 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97zf7\" (UniqueName: \"kubernetes.io/projected/c3808ad8-b638-412a-b668-00d8e0e76528-kube-api-access-97zf7\") on node \"crc\" DevicePath \"\"" Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.796993 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3808ad8-b638-412a-b668-00d8e0e76528-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3808ad8-b638-412a-b668-00d8e0e76528" (UID: "c3808ad8-b638-412a-b668-00d8e0e76528"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.895826 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3808ad8-b638-412a-b668-00d8e0e76528-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.919503 4857 generic.go:334] "Generic (PLEG): container finished" podID="c3808ad8-b638-412a-b668-00d8e0e76528" containerID="9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a" exitCode=0 Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.919567 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn8cm" event={"ID":"c3808ad8-b638-412a-b668-00d8e0e76528","Type":"ContainerDied","Data":"9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a"} Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.919588 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn8cm" Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.919615 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn8cm" event={"ID":"c3808ad8-b638-412a-b668-00d8e0e76528","Type":"ContainerDied","Data":"c325c3c2004ca0ad6999610dc1303f005eb1bd15eb42e619ca5f58afc03921f1"} Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.919645 4857 scope.go:117] "RemoveContainer" containerID="9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a" Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.966671 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qn8cm"] Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.966818 4857 scope.go:117] "RemoveContainer" containerID="0b4cb7884b3af24aaaa58e7b361b641d080b5beb57b38287f795e6c3ce20cf69" Dec 01 22:50:19 crc kubenswrapper[4857]: I1201 22:50:19.975332 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qn8cm"] Dec 01 22:50:20 crc kubenswrapper[4857]: I1201 22:50:20.029379 4857 scope.go:117] "RemoveContainer" containerID="f3d3ac39225d449559d773a7399292029402199d14ef78eb458b64472f5d2178" Dec 01 22:50:20 crc kubenswrapper[4857]: I1201 22:50:20.053895 4857 scope.go:117] "RemoveContainer" containerID="9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a" Dec 01 22:50:20 crc kubenswrapper[4857]: E1201 22:50:20.054336 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a\": container with ID starting with 9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a not found: ID does not exist" containerID="9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a" Dec 01 22:50:20 crc kubenswrapper[4857]: I1201 22:50:20.054370 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a"} err="failed to get container status \"9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a\": rpc error: code = NotFound desc = could not find container \"9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a\": container with ID starting with 9ba33701299a768d914eb2b71379fe85edb05b27b000ead047e49b310703945a not found: ID does not exist" Dec 01 22:50:20 crc kubenswrapper[4857]: I1201 22:50:20.054394 4857 scope.go:117] "RemoveContainer" containerID="0b4cb7884b3af24aaaa58e7b361b641d080b5beb57b38287f795e6c3ce20cf69" Dec 01 22:50:20 crc kubenswrapper[4857]: E1201 22:50:20.054849 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b4cb7884b3af24aaaa58e7b361b641d080b5beb57b38287f795e6c3ce20cf69\": container with ID starting with 0b4cb7884b3af24aaaa58e7b361b641d080b5beb57b38287f795e6c3ce20cf69 not found: ID does not exist" containerID="0b4cb7884b3af24aaaa58e7b361b641d080b5beb57b38287f795e6c3ce20cf69" Dec 01 22:50:20 crc kubenswrapper[4857]: I1201 22:50:20.054894 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b4cb7884b3af24aaaa58e7b361b641d080b5beb57b38287f795e6c3ce20cf69"} err="failed to get container status \"0b4cb7884b3af24aaaa58e7b361b641d080b5beb57b38287f795e6c3ce20cf69\": rpc error: code = NotFound desc = could not find container \"0b4cb7884b3af24aaaa58e7b361b641d080b5beb57b38287f795e6c3ce20cf69\": container with ID starting with 0b4cb7884b3af24aaaa58e7b361b641d080b5beb57b38287f795e6c3ce20cf69 not found: ID does not exist" Dec 01 22:50:20 crc kubenswrapper[4857]: I1201 22:50:20.054958 4857 scope.go:117] "RemoveContainer" containerID="f3d3ac39225d449559d773a7399292029402199d14ef78eb458b64472f5d2178" Dec 01 22:50:20 crc kubenswrapper[4857]: E1201 22:50:20.055563 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3d3ac39225d449559d773a7399292029402199d14ef78eb458b64472f5d2178\": container with ID starting with f3d3ac39225d449559d773a7399292029402199d14ef78eb458b64472f5d2178 not found: ID does not exist" containerID="f3d3ac39225d449559d773a7399292029402199d14ef78eb458b64472f5d2178" Dec 01 22:50:20 crc kubenswrapper[4857]: I1201 22:50:20.055592 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3d3ac39225d449559d773a7399292029402199d14ef78eb458b64472f5d2178"} err="failed to get container status \"f3d3ac39225d449559d773a7399292029402199d14ef78eb458b64472f5d2178\": rpc error: code = NotFound desc = could not find container \"f3d3ac39225d449559d773a7399292029402199d14ef78eb458b64472f5d2178\": container with ID starting with f3d3ac39225d449559d773a7399292029402199d14ef78eb458b64472f5d2178 not found: ID does not exist" Dec 01 22:50:21 crc kubenswrapper[4857]: I1201 22:50:21.845778 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3808ad8-b638-412a-b668-00d8e0e76528" path="/var/lib/kubelet/pods/c3808ad8-b638-412a-b668-00d8e0e76528/volumes" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.571940 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qwhq9/must-gather-h5lcq"] Dec 01 22:50:23 crc kubenswrapper[4857]: E1201 22:50:23.572667 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3808ad8-b638-412a-b668-00d8e0e76528" containerName="extract-utilities" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.572681 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3808ad8-b638-412a-b668-00d8e0e76528" containerName="extract-utilities" Dec 01 22:50:23 crc kubenswrapper[4857]: E1201 22:50:23.572696 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3808ad8-b638-412a-b668-00d8e0e76528" containerName="registry-server" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.572702 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3808ad8-b638-412a-b668-00d8e0e76528" containerName="registry-server" Dec 01 22:50:23 crc kubenswrapper[4857]: E1201 22:50:23.572718 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3808ad8-b638-412a-b668-00d8e0e76528" containerName="extract-content" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.572726 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3808ad8-b638-412a-b668-00d8e0e76528" containerName="extract-content" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.572937 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3808ad8-b638-412a-b668-00d8e0e76528" containerName="registry-server" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.574239 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qwhq9/must-gather-h5lcq" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.583614 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qwhq9"/"kube-root-ca.crt" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.583804 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qwhq9"/"openshift-service-ca.crt" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.589426 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qwhq9/must-gather-h5lcq"] Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.705686 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/51a7c54e-a7cb-49c7-83bc-621241a2dd1a-must-gather-output\") pod \"must-gather-h5lcq\" (UID: \"51a7c54e-a7cb-49c7-83bc-621241a2dd1a\") " pod="openshift-must-gather-qwhq9/must-gather-h5lcq" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.706087 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbzv7\" (UniqueName: \"kubernetes.io/projected/51a7c54e-a7cb-49c7-83bc-621241a2dd1a-kube-api-access-jbzv7\") pod \"must-gather-h5lcq\" (UID: \"51a7c54e-a7cb-49c7-83bc-621241a2dd1a\") " pod="openshift-must-gather-qwhq9/must-gather-h5lcq" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.808162 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/51a7c54e-a7cb-49c7-83bc-621241a2dd1a-must-gather-output\") pod \"must-gather-h5lcq\" (UID: \"51a7c54e-a7cb-49c7-83bc-621241a2dd1a\") " pod="openshift-must-gather-qwhq9/must-gather-h5lcq" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.808242 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbzv7\" (UniqueName: \"kubernetes.io/projected/51a7c54e-a7cb-49c7-83bc-621241a2dd1a-kube-api-access-jbzv7\") pod \"must-gather-h5lcq\" (UID: \"51a7c54e-a7cb-49c7-83bc-621241a2dd1a\") " pod="openshift-must-gather-qwhq9/must-gather-h5lcq" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.808895 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/51a7c54e-a7cb-49c7-83bc-621241a2dd1a-must-gather-output\") pod \"must-gather-h5lcq\" (UID: \"51a7c54e-a7cb-49c7-83bc-621241a2dd1a\") " pod="openshift-must-gather-qwhq9/must-gather-h5lcq" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.836230 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbzv7\" (UniqueName: \"kubernetes.io/projected/51a7c54e-a7cb-49c7-83bc-621241a2dd1a-kube-api-access-jbzv7\") pod \"must-gather-h5lcq\" (UID: \"51a7c54e-a7cb-49c7-83bc-621241a2dd1a\") " pod="openshift-must-gather-qwhq9/must-gather-h5lcq" Dec 01 22:50:23 crc kubenswrapper[4857]: I1201 22:50:23.893483 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qwhq9/must-gather-h5lcq" Dec 01 22:50:24 crc kubenswrapper[4857]: I1201 22:50:24.430710 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qwhq9/must-gather-h5lcq"] Dec 01 22:50:24 crc kubenswrapper[4857]: W1201 22:50:24.432947 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51a7c54e_a7cb_49c7_83bc_621241a2dd1a.slice/crio-4bc1151ac2388cef601365c4ebb38ddbb32202e9cac73f017c6354f4a331fc2e WatchSource:0}: Error finding container 4bc1151ac2388cef601365c4ebb38ddbb32202e9cac73f017c6354f4a331fc2e: Status 404 returned error can't find the container with id 4bc1151ac2388cef601365c4ebb38ddbb32202e9cac73f017c6354f4a331fc2e Dec 01 22:50:24 crc kubenswrapper[4857]: I1201 22:50:24.983601 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qwhq9/must-gather-h5lcq" event={"ID":"51a7c54e-a7cb-49c7-83bc-621241a2dd1a","Type":"ContainerStarted","Data":"c9424a52371cf8f1c12d017bed32fe71f093641cb58f2bbdaaca0072bca7c936"} Dec 01 22:50:24 crc kubenswrapper[4857]: I1201 22:50:24.983875 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qwhq9/must-gather-h5lcq" event={"ID":"51a7c54e-a7cb-49c7-83bc-621241a2dd1a","Type":"ContainerStarted","Data":"a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60"} Dec 01 22:50:24 crc kubenswrapper[4857]: I1201 22:50:24.983885 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qwhq9/must-gather-h5lcq" event={"ID":"51a7c54e-a7cb-49c7-83bc-621241a2dd1a","Type":"ContainerStarted","Data":"4bc1151ac2388cef601365c4ebb38ddbb32202e9cac73f017c6354f4a331fc2e"} Dec 01 22:50:25 crc kubenswrapper[4857]: I1201 22:50:25.007762 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qwhq9/must-gather-h5lcq" podStartSLOduration=2.007739392 podStartE2EDuration="2.007739392s" podCreationTimestamp="2025-12-01 22:50:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 22:50:25.002030002 +0000 UTC m=+4543.492092339" watchObservedRunningTime="2025-12-01 22:50:25.007739392 +0000 UTC m=+4543.497801719" Dec 01 22:50:27 crc kubenswrapper[4857]: I1201 22:50:27.805714 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:50:27 crc kubenswrapper[4857]: I1201 22:50:27.806245 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:50:29 crc kubenswrapper[4857]: I1201 22:50:29.009278 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qwhq9/crc-debug-9bg9z"] Dec 01 22:50:29 crc kubenswrapper[4857]: I1201 22:50:29.010988 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" Dec 01 22:50:29 crc kubenswrapper[4857]: I1201 22:50:29.013223 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qwhq9"/"default-dockercfg-xfjrq" Dec 01 22:50:29 crc kubenswrapper[4857]: I1201 22:50:29.032720 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vll2f\" (UniqueName: \"kubernetes.io/projected/1331ba74-3c6a-4af2-8720-c5d0d0b30223-kube-api-access-vll2f\") pod \"crc-debug-9bg9z\" (UID: \"1331ba74-3c6a-4af2-8720-c5d0d0b30223\") " pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" Dec 01 22:50:29 crc kubenswrapper[4857]: I1201 22:50:29.034585 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1331ba74-3c6a-4af2-8720-c5d0d0b30223-host\") pod \"crc-debug-9bg9z\" (UID: \"1331ba74-3c6a-4af2-8720-c5d0d0b30223\") " pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" Dec 01 22:50:29 crc kubenswrapper[4857]: I1201 22:50:29.135464 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1331ba74-3c6a-4af2-8720-c5d0d0b30223-host\") pod \"crc-debug-9bg9z\" (UID: \"1331ba74-3c6a-4af2-8720-c5d0d0b30223\") " pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" Dec 01 22:50:29 crc kubenswrapper[4857]: I1201 22:50:29.135610 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vll2f\" (UniqueName: \"kubernetes.io/projected/1331ba74-3c6a-4af2-8720-c5d0d0b30223-kube-api-access-vll2f\") pod \"crc-debug-9bg9z\" (UID: \"1331ba74-3c6a-4af2-8720-c5d0d0b30223\") " pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" Dec 01 22:50:29 crc kubenswrapper[4857]: I1201 22:50:29.136106 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1331ba74-3c6a-4af2-8720-c5d0d0b30223-host\") pod \"crc-debug-9bg9z\" (UID: \"1331ba74-3c6a-4af2-8720-c5d0d0b30223\") " pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" Dec 01 22:50:29 crc kubenswrapper[4857]: I1201 22:50:29.162055 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vll2f\" (UniqueName: \"kubernetes.io/projected/1331ba74-3c6a-4af2-8720-c5d0d0b30223-kube-api-access-vll2f\") pod \"crc-debug-9bg9z\" (UID: \"1331ba74-3c6a-4af2-8720-c5d0d0b30223\") " pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" Dec 01 22:50:29 crc kubenswrapper[4857]: I1201 22:50:29.332236 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" Dec 01 22:50:29 crc kubenswrapper[4857]: W1201 22:50:29.368468 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1331ba74_3c6a_4af2_8720_c5d0d0b30223.slice/crio-3411a9ac3bea3ba9a7805093b541345ebd68d9e4592f71d3f174630a89629ddb WatchSource:0}: Error finding container 3411a9ac3bea3ba9a7805093b541345ebd68d9e4592f71d3f174630a89629ddb: Status 404 returned error can't find the container with id 3411a9ac3bea3ba9a7805093b541345ebd68d9e4592f71d3f174630a89629ddb Dec 01 22:50:30 crc kubenswrapper[4857]: I1201 22:50:30.038328 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" event={"ID":"1331ba74-3c6a-4af2-8720-c5d0d0b30223","Type":"ContainerStarted","Data":"72ca8cf9dd95baf34315474e244185177696ea6c0b3f132fd96b4bfe643a94f0"} Dec 01 22:50:30 crc kubenswrapper[4857]: I1201 22:50:30.038841 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" event={"ID":"1331ba74-3c6a-4af2-8720-c5d0d0b30223","Type":"ContainerStarted","Data":"3411a9ac3bea3ba9a7805093b541345ebd68d9e4592f71d3f174630a89629ddb"} Dec 01 22:50:30 crc kubenswrapper[4857]: I1201 22:50:30.067667 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" podStartSLOduration=2.067646684 podStartE2EDuration="2.067646684s" podCreationTimestamp="2025-12-01 22:50:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 22:50:30.065963423 +0000 UTC m=+4548.556025740" watchObservedRunningTime="2025-12-01 22:50:30.067646684 +0000 UTC m=+4548.557709001" Dec 01 22:50:41 crc kubenswrapper[4857]: I1201 22:50:41.167712 4857 generic.go:334] "Generic (PLEG): container finished" podID="1331ba74-3c6a-4af2-8720-c5d0d0b30223" containerID="72ca8cf9dd95baf34315474e244185177696ea6c0b3f132fd96b4bfe643a94f0" exitCode=0 Dec 01 22:50:41 crc kubenswrapper[4857]: I1201 22:50:41.167797 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" event={"ID":"1331ba74-3c6a-4af2-8720-c5d0d0b30223","Type":"ContainerDied","Data":"72ca8cf9dd95baf34315474e244185177696ea6c0b3f132fd96b4bfe643a94f0"} Dec 01 22:50:42 crc kubenswrapper[4857]: I1201 22:50:42.317766 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" Dec 01 22:50:42 crc kubenswrapper[4857]: I1201 22:50:42.349391 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qwhq9/crc-debug-9bg9z"] Dec 01 22:50:42 crc kubenswrapper[4857]: I1201 22:50:42.360656 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qwhq9/crc-debug-9bg9z"] Dec 01 22:50:42 crc kubenswrapper[4857]: I1201 22:50:42.362962 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1331ba74-3c6a-4af2-8720-c5d0d0b30223-host\") pod \"1331ba74-3c6a-4af2-8720-c5d0d0b30223\" (UID: \"1331ba74-3c6a-4af2-8720-c5d0d0b30223\") " Dec 01 22:50:42 crc kubenswrapper[4857]: I1201 22:50:42.363052 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1331ba74-3c6a-4af2-8720-c5d0d0b30223-host" (OuterVolumeSpecName: "host") pod "1331ba74-3c6a-4af2-8720-c5d0d0b30223" (UID: "1331ba74-3c6a-4af2-8720-c5d0d0b30223"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 22:50:42 crc kubenswrapper[4857]: I1201 22:50:42.363126 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vll2f\" (UniqueName: \"kubernetes.io/projected/1331ba74-3c6a-4af2-8720-c5d0d0b30223-kube-api-access-vll2f\") pod \"1331ba74-3c6a-4af2-8720-c5d0d0b30223\" (UID: \"1331ba74-3c6a-4af2-8720-c5d0d0b30223\") " Dec 01 22:50:42 crc kubenswrapper[4857]: I1201 22:50:42.363685 4857 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1331ba74-3c6a-4af2-8720-c5d0d0b30223-host\") on node \"crc\" DevicePath \"\"" Dec 01 22:50:42 crc kubenswrapper[4857]: I1201 22:50:42.376719 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1331ba74-3c6a-4af2-8720-c5d0d0b30223-kube-api-access-vll2f" (OuterVolumeSpecName: "kube-api-access-vll2f") pod "1331ba74-3c6a-4af2-8720-c5d0d0b30223" (UID: "1331ba74-3c6a-4af2-8720-c5d0d0b30223"). InnerVolumeSpecName "kube-api-access-vll2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:50:42 crc kubenswrapper[4857]: I1201 22:50:42.465371 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vll2f\" (UniqueName: \"kubernetes.io/projected/1331ba74-3c6a-4af2-8720-c5d0d0b30223-kube-api-access-vll2f\") on node \"crc\" DevicePath \"\"" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.188568 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3411a9ac3bea3ba9a7805093b541345ebd68d9e4592f71d3f174630a89629ddb" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.188640 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qwhq9/crc-debug-9bg9z" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.581178 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qwhq9/crc-debug-v2zgh"] Dec 01 22:50:43 crc kubenswrapper[4857]: E1201 22:50:43.581546 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1331ba74-3c6a-4af2-8720-c5d0d0b30223" containerName="container-00" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.581558 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="1331ba74-3c6a-4af2-8720-c5d0d0b30223" containerName="container-00" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.581771 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="1331ba74-3c6a-4af2-8720-c5d0d0b30223" containerName="container-00" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.582453 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qwhq9/crc-debug-v2zgh" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.587216 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qwhq9"/"default-dockercfg-xfjrq" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.603825 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd4cd826-888f-40c3-a91c-8982febb154d-host\") pod \"crc-debug-v2zgh\" (UID: \"bd4cd826-888f-40c3-a91c-8982febb154d\") " pod="openshift-must-gather-qwhq9/crc-debug-v2zgh" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.603928 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmqqs\" (UniqueName: \"kubernetes.io/projected/bd4cd826-888f-40c3-a91c-8982febb154d-kube-api-access-hmqqs\") pod \"crc-debug-v2zgh\" (UID: \"bd4cd826-888f-40c3-a91c-8982febb154d\") " pod="openshift-must-gather-qwhq9/crc-debug-v2zgh" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.706459 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd4cd826-888f-40c3-a91c-8982febb154d-host\") pod \"crc-debug-v2zgh\" (UID: \"bd4cd826-888f-40c3-a91c-8982febb154d\") " pod="openshift-must-gather-qwhq9/crc-debug-v2zgh" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.706626 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd4cd826-888f-40c3-a91c-8982febb154d-host\") pod \"crc-debug-v2zgh\" (UID: \"bd4cd826-888f-40c3-a91c-8982febb154d\") " pod="openshift-must-gather-qwhq9/crc-debug-v2zgh" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.706644 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmqqs\" (UniqueName: \"kubernetes.io/projected/bd4cd826-888f-40c3-a91c-8982febb154d-kube-api-access-hmqqs\") pod \"crc-debug-v2zgh\" (UID: \"bd4cd826-888f-40c3-a91c-8982febb154d\") " pod="openshift-must-gather-qwhq9/crc-debug-v2zgh" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.732819 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmqqs\" (UniqueName: \"kubernetes.io/projected/bd4cd826-888f-40c3-a91c-8982febb154d-kube-api-access-hmqqs\") pod \"crc-debug-v2zgh\" (UID: \"bd4cd826-888f-40c3-a91c-8982febb154d\") " pod="openshift-must-gather-qwhq9/crc-debug-v2zgh" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.845538 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1331ba74-3c6a-4af2-8720-c5d0d0b30223" path="/var/lib/kubelet/pods/1331ba74-3c6a-4af2-8720-c5d0d0b30223/volumes" Dec 01 22:50:43 crc kubenswrapper[4857]: I1201 22:50:43.907468 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qwhq9/crc-debug-v2zgh" Dec 01 22:50:43 crc kubenswrapper[4857]: W1201 22:50:43.940945 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd4cd826_888f_40c3_a91c_8982febb154d.slice/crio-82f945282a00b0ee894e6d2c80cb5275d2dccca48182cc02466559d429aa38be WatchSource:0}: Error finding container 82f945282a00b0ee894e6d2c80cb5275d2dccca48182cc02466559d429aa38be: Status 404 returned error can't find the container with id 82f945282a00b0ee894e6d2c80cb5275d2dccca48182cc02466559d429aa38be Dec 01 22:50:44 crc kubenswrapper[4857]: I1201 22:50:44.198606 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qwhq9/crc-debug-v2zgh" event={"ID":"bd4cd826-888f-40c3-a91c-8982febb154d","Type":"ContainerStarted","Data":"82f945282a00b0ee894e6d2c80cb5275d2dccca48182cc02466559d429aa38be"} Dec 01 22:50:45 crc kubenswrapper[4857]: I1201 22:50:45.211727 4857 generic.go:334] "Generic (PLEG): container finished" podID="bd4cd826-888f-40c3-a91c-8982febb154d" containerID="69fc2fbef2473ae861a136947d271692b466df351376c9d77f3b0f461982fa17" exitCode=1 Dec 01 22:50:45 crc kubenswrapper[4857]: I1201 22:50:45.211770 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qwhq9/crc-debug-v2zgh" event={"ID":"bd4cd826-888f-40c3-a91c-8982febb154d","Type":"ContainerDied","Data":"69fc2fbef2473ae861a136947d271692b466df351376c9d77f3b0f461982fa17"} Dec 01 22:50:45 crc kubenswrapper[4857]: I1201 22:50:45.257131 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qwhq9/crc-debug-v2zgh"] Dec 01 22:50:45 crc kubenswrapper[4857]: I1201 22:50:45.269651 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qwhq9/crc-debug-v2zgh"] Dec 01 22:50:46 crc kubenswrapper[4857]: I1201 22:50:46.348722 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qwhq9/crc-debug-v2zgh" Dec 01 22:50:46 crc kubenswrapper[4857]: I1201 22:50:46.378105 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmqqs\" (UniqueName: \"kubernetes.io/projected/bd4cd826-888f-40c3-a91c-8982febb154d-kube-api-access-hmqqs\") pod \"bd4cd826-888f-40c3-a91c-8982febb154d\" (UID: \"bd4cd826-888f-40c3-a91c-8982febb154d\") " Dec 01 22:50:46 crc kubenswrapper[4857]: I1201 22:50:46.378253 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd4cd826-888f-40c3-a91c-8982febb154d-host\") pod \"bd4cd826-888f-40c3-a91c-8982febb154d\" (UID: \"bd4cd826-888f-40c3-a91c-8982febb154d\") " Dec 01 22:50:46 crc kubenswrapper[4857]: I1201 22:50:46.378390 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd4cd826-888f-40c3-a91c-8982febb154d-host" (OuterVolumeSpecName: "host") pod "bd4cd826-888f-40c3-a91c-8982febb154d" (UID: "bd4cd826-888f-40c3-a91c-8982febb154d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 22:50:46 crc kubenswrapper[4857]: I1201 22:50:46.378888 4857 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd4cd826-888f-40c3-a91c-8982febb154d-host\") on node \"crc\" DevicePath \"\"" Dec 01 22:50:46 crc kubenswrapper[4857]: I1201 22:50:46.383467 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd4cd826-888f-40c3-a91c-8982febb154d-kube-api-access-hmqqs" (OuterVolumeSpecName: "kube-api-access-hmqqs") pod "bd4cd826-888f-40c3-a91c-8982febb154d" (UID: "bd4cd826-888f-40c3-a91c-8982febb154d"). InnerVolumeSpecName "kube-api-access-hmqqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:50:46 crc kubenswrapper[4857]: I1201 22:50:46.480607 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmqqs\" (UniqueName: \"kubernetes.io/projected/bd4cd826-888f-40c3-a91c-8982febb154d-kube-api-access-hmqqs\") on node \"crc\" DevicePath \"\"" Dec 01 22:50:47 crc kubenswrapper[4857]: I1201 22:50:47.232233 4857 scope.go:117] "RemoveContainer" containerID="69fc2fbef2473ae861a136947d271692b466df351376c9d77f3b0f461982fa17" Dec 01 22:50:47 crc kubenswrapper[4857]: I1201 22:50:47.232327 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qwhq9/crc-debug-v2zgh" Dec 01 22:50:47 crc kubenswrapper[4857]: I1201 22:50:47.846444 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd4cd826-888f-40c3-a91c-8982febb154d" path="/var/lib/kubelet/pods/bd4cd826-888f-40c3-a91c-8982febb154d/volumes" Dec 01 22:50:57 crc kubenswrapper[4857]: I1201 22:50:57.806707 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:50:57 crc kubenswrapper[4857]: I1201 22:50:57.807272 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:50:57 crc kubenswrapper[4857]: I1201 22:50:57.807328 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 22:50:57 crc kubenswrapper[4857]: I1201 22:50:57.808188 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 22:50:57 crc kubenswrapper[4857]: I1201 22:50:57.808249 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" gracePeriod=600 Dec 01 22:50:58 crc kubenswrapper[4857]: E1201 22:50:58.074930 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:50:58 crc kubenswrapper[4857]: I1201 22:50:58.360578 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" exitCode=0 Dec 01 22:50:58 crc kubenswrapper[4857]: I1201 22:50:58.360646 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4"} Dec 01 22:50:58 crc kubenswrapper[4857]: I1201 22:50:58.360755 4857 scope.go:117] "RemoveContainer" containerID="4d908912b33495ea153b8b1fbe5c79474599457200a6669673854ac4fb373bcb" Dec 01 22:50:58 crc kubenswrapper[4857]: I1201 22:50:58.361664 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:50:58 crc kubenswrapper[4857]: E1201 22:50:58.362105 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:51:11 crc kubenswrapper[4857]: I1201 22:51:11.841845 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:51:11 crc kubenswrapper[4857]: E1201 22:51:11.843801 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:51:24 crc kubenswrapper[4857]: I1201 22:51:24.836429 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:51:24 crc kubenswrapper[4857]: E1201 22:51:24.837513 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:51:39 crc kubenswrapper[4857]: I1201 22:51:39.846887 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:51:39 crc kubenswrapper[4857]: E1201 22:51:39.847820 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:51:52 crc kubenswrapper[4857]: I1201 22:51:52.835243 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:51:52 crc kubenswrapper[4857]: E1201 22:51:52.836177 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:51:57 crc kubenswrapper[4857]: I1201 22:51:57.414198 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_98dbb0ff-4644-431e-801e-5ce7fb3f6f81/init-config-reloader/0.log" Dec 01 22:51:57 crc kubenswrapper[4857]: I1201 22:51:57.574476 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_98dbb0ff-4644-431e-801e-5ce7fb3f6f81/init-config-reloader/0.log" Dec 01 22:51:57 crc kubenswrapper[4857]: I1201 22:51:57.594462 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_98dbb0ff-4644-431e-801e-5ce7fb3f6f81/alertmanager/0.log" Dec 01 22:51:57 crc kubenswrapper[4857]: I1201 22:51:57.645822 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_98dbb0ff-4644-431e-801e-5ce7fb3f6f81/config-reloader/0.log" Dec 01 22:51:57 crc kubenswrapper[4857]: I1201 22:51:57.743110 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e78cc157-06d0-4c5f-8da8-6e1a9591600a/aodh-api/0.log" Dec 01 22:51:57 crc kubenswrapper[4857]: I1201 22:51:57.767274 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e78cc157-06d0-4c5f-8da8-6e1a9591600a/aodh-evaluator/0.log" Dec 01 22:51:57 crc kubenswrapper[4857]: I1201 22:51:57.878797 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e78cc157-06d0-4c5f-8da8-6e1a9591600a/aodh-notifier/0.log" Dec 01 22:51:57 crc kubenswrapper[4857]: I1201 22:51:57.880666 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e78cc157-06d0-4c5f-8da8-6e1a9591600a/aodh-listener/0.log" Dec 01 22:51:57 crc kubenswrapper[4857]: I1201 22:51:57.926510 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-cf44fcf6b-c86cx_13134eb9-c4d1-4060-8fd3-266afd46f420/barbican-api/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.069283 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-cf44fcf6b-c86cx_13134eb9-c4d1-4060-8fd3-266afd46f420/barbican-api-log/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.138536 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-db75d4bb-tjp4b_d92f4970-9829-4417-9cc1-9436c8993a18/barbican-keystone-listener/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.168694 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-db75d4bb-tjp4b_d92f4970-9829-4417-9cc1-9436c8993a18/barbican-keystone-listener-log/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.274425 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5f6bc6688c-77x7p_5ae783f5-078d-49c0-9b57-cf796b0e8f3b/barbican-worker/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.318389 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5f6bc6688c-77x7p_5ae783f5-078d-49c0-9b57-cf796b0e8f3b/barbican-worker-log/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.469288 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k9mwt_3916d4da-634b-4ade-be14-3a0c60e6a9d2/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.493229 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ccfb3129-79e9-4397-93c3-05ddbd0e6f76/ceilometer-central-agent/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.610895 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ccfb3129-79e9-4397-93c3-05ddbd0e6f76/ceilometer-notification-agent/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.656840 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ccfb3129-79e9-4397-93c3-05ddbd0e6f76/proxy-httpd/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.725502 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ccfb3129-79e9-4397-93c3-05ddbd0e6f76/sg-core/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.826062 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_919f0ad8-05ea-44a1-a4ae-f81f4255e8da/cinder-api/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.932879 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_919f0ad8-05ea-44a1-a4ae-f81f4255e8da/cinder-api-log/0.log" Dec 01 22:51:58 crc kubenswrapper[4857]: I1201 22:51:58.993347 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f38eac29-d14a-4bf8-bbd8-df39c591ee6a/cinder-scheduler/0.log" Dec 01 22:51:59 crc kubenswrapper[4857]: I1201 22:51:59.046061 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f38eac29-d14a-4bf8-bbd8-df39c591ee6a/probe/0.log" Dec 01 22:51:59 crc kubenswrapper[4857]: I1201 22:51:59.188291 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wn4rm_0a768b62-44a4-4493-bd0d-2a536c55370d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:51:59 crc kubenswrapper[4857]: I1201 22:51:59.275662 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-7mccz_e3196a47-b0ed-4f5d-a639-496ff78d1acd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:51:59 crc kubenswrapper[4857]: I1201 22:51:59.383223 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-5mp7g_72ba8edd-5f69-4c93-9954-aaa52e18c2fa/init/0.log" Dec 01 22:51:59 crc kubenswrapper[4857]: I1201 22:51:59.520747 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-5mp7g_72ba8edd-5f69-4c93-9954-aaa52e18c2fa/init/0.log" Dec 01 22:51:59 crc kubenswrapper[4857]: I1201 22:51:59.577682 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-5mp7g_72ba8edd-5f69-4c93-9954-aaa52e18c2fa/dnsmasq-dns/0.log" Dec 01 22:51:59 crc kubenswrapper[4857]: I1201 22:51:59.612911 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-ltnd2_d419ce3b-db4a-4224-a322-69aaa24650fa/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:51:59 crc kubenswrapper[4857]: I1201 22:51:59.778893 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_dc169c0b-67f2-48bf-8ac6-f58ca383ae4c/glance-httpd/0.log" Dec 01 22:51:59 crc kubenswrapper[4857]: I1201 22:51:59.785879 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_dc169c0b-67f2-48bf-8ac6-f58ca383ae4c/glance-log/0.log" Dec 01 22:51:59 crc kubenswrapper[4857]: I1201 22:51:59.939850 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4f3b7eba-7e44-4f50-a2da-626cfe8e3921/glance-log/0.log" Dec 01 22:51:59 crc kubenswrapper[4857]: I1201 22:51:59.995215 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4f3b7eba-7e44-4f50-a2da-626cfe8e3921/glance-httpd/0.log" Dec 01 22:52:00 crc kubenswrapper[4857]: I1201 22:52:00.445070 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-5c86f5f679-g6dgl_05d88f39-b756-4b31-9385-99213e839003/heat-api/0.log" Dec 01 22:52:00 crc kubenswrapper[4857]: I1201 22:52:00.752591 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-7c64685c4b-rgnfp_d9fd2d2f-01df-4aa4-bc83-46f30fb0c592/heat-engine/0.log" Dec 01 22:52:01 crc kubenswrapper[4857]: I1201 22:52:01.012202 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-gfgvp_f71f02cd-2132-437a-b18c-e3513eb0710c/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:52:01 crc kubenswrapper[4857]: I1201 22:52:01.053537 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5894b69764-fpz5r_ce4b4968-ffc0-4d0d-86b5-da72a0712630/heat-cfnapi/0.log" Dec 01 22:52:01 crc kubenswrapper[4857]: I1201 22:52:01.076406 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-2zzdx_696c86fa-0abd-4bae-9821-f570a70ad6e0/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:52:01 crc kubenswrapper[4857]: I1201 22:52:01.181600 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29410441-m46zq_a8480d29-0c86-4ee6-85ea-6be3087135ec/keystone-cron/0.log" Dec 01 22:52:01 crc kubenswrapper[4857]: I1201 22:52:01.310253 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_cfbeeaca-f2fe-41ed-bb4d-092cb86751d5/kube-state-metrics/0.log" Dec 01 22:52:01 crc kubenswrapper[4857]: I1201 22:52:01.356553 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-fdd4494c7-87llf_085e66d4-f84a-47e2-851b-99f08ccc9683/keystone-api/0.log" Dec 01 22:52:01 crc kubenswrapper[4857]: I1201 22:52:01.467448 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-vrkv4_afe24796-7855-411d-b62a-a4e5fefc12c6/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:52:02 crc kubenswrapper[4857]: I1201 22:52:02.197516 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pd9x_60f230d9-9293-4ab7-a2c5-5a9e3c5dd623/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:52:02 crc kubenswrapper[4857]: I1201 22:52:02.241966 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57b8cf5fc7-kg4kw_33d2facf-f0c8-4566-a15c-038f0f9f9a6e/neutron-httpd/0.log" Dec 01 22:52:02 crc kubenswrapper[4857]: I1201 22:52:02.280134 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57b8cf5fc7-kg4kw_33d2facf-f0c8-4566-a15c-038f0f9f9a6e/neutron-api/0.log" Dec 01 22:52:02 crc kubenswrapper[4857]: I1201 22:52:02.654109 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_dcd76a72-5869-4014-8bf7-ee57297feb85/nova-api-log/0.log" Dec 01 22:52:02 crc kubenswrapper[4857]: I1201 22:52:02.906535 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a69a1083-f915-4820-82f3-f82c5053d365/nova-cell0-conductor-conductor/0.log" Dec 01 22:52:03 crc kubenswrapper[4857]: I1201 22:52:03.122923 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_dcd76a72-5869-4014-8bf7-ee57297feb85/nova-api-api/0.log" Dec 01 22:52:03 crc kubenswrapper[4857]: I1201 22:52:03.188702 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_cd0a02ec-0eec-4459-9de2-96f044ff45c7/nova-cell1-conductor-conductor/0.log" Dec 01 22:52:03 crc kubenswrapper[4857]: I1201 22:52:03.325961 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e28cb1b9-5dff-4741-ac99-f72b10c0c09d/nova-cell1-novncproxy-novncproxy/0.log" Dec 01 22:52:03 crc kubenswrapper[4857]: I1201 22:52:03.390463 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-l7ltc_3b8d39f4-2804-4fda-bec9-0cd4c2401f42/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:52:03 crc kubenswrapper[4857]: I1201 22:52:03.741298 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f3a844a9-6868-43b6-b5a9-402e619f0d50/nova-metadata-log/0.log" Dec 01 22:52:03 crc kubenswrapper[4857]: I1201 22:52:03.838030 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:52:03 crc kubenswrapper[4857]: E1201 22:52:03.838303 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:52:03 crc kubenswrapper[4857]: I1201 22:52:03.944587 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_2859495a-e026-400f-a304-05e25e961f2e/nova-scheduler-scheduler/0.log" Dec 01 22:52:03 crc kubenswrapper[4857]: I1201 22:52:03.967397 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_be0a242c-3c02-4d1e-910b-21e74e7c4e43/mysql-bootstrap/0.log" Dec 01 22:52:04 crc kubenswrapper[4857]: I1201 22:52:04.176411 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_be0a242c-3c02-4d1e-910b-21e74e7c4e43/mysql-bootstrap/0.log" Dec 01 22:52:04 crc kubenswrapper[4857]: I1201 22:52:04.256698 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_be0a242c-3c02-4d1e-910b-21e74e7c4e43/galera/0.log" Dec 01 22:52:04 crc kubenswrapper[4857]: I1201 22:52:04.422775 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b3755be2-a089-4259-af43-6e0fbedef24b/mysql-bootstrap/0.log" Dec 01 22:52:04 crc kubenswrapper[4857]: I1201 22:52:04.556826 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b3755be2-a089-4259-af43-6e0fbedef24b/mysql-bootstrap/0.log" Dec 01 22:52:04 crc kubenswrapper[4857]: I1201 22:52:04.576009 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b3755be2-a089-4259-af43-6e0fbedef24b/galera/0.log" Dec 01 22:52:04 crc kubenswrapper[4857]: I1201 22:52:04.735094 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_8b715e70-6964-4dd3-9c88-67d41fd7ad7a/openstackclient/0.log" Dec 01 22:52:04 crc kubenswrapper[4857]: I1201 22:52:04.876874 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lqjrc_7aab234d-a569-48bb-b75b-6a4ecf993df7/openstack-network-exporter/0.log" Dec 01 22:52:05 crc kubenswrapper[4857]: I1201 22:52:05.056070 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-nhfhn_32213ad3-4287-4968-9d42-bfeb2ff558d1/ovn-controller/0.log" Dec 01 22:52:05 crc kubenswrapper[4857]: I1201 22:52:05.125218 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f3a844a9-6868-43b6-b5a9-402e619f0d50/nova-metadata-metadata/0.log" Dec 01 22:52:05 crc kubenswrapper[4857]: I1201 22:52:05.213601 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gwbmr_5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f/ovsdb-server-init/0.log" Dec 01 22:52:05 crc kubenswrapper[4857]: I1201 22:52:05.433928 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gwbmr_5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f/ovsdb-server-init/0.log" Dec 01 22:52:05 crc kubenswrapper[4857]: I1201 22:52:05.457201 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gwbmr_5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f/ovs-vswitchd/0.log" Dec 01 22:52:05 crc kubenswrapper[4857]: I1201 22:52:05.464512 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gwbmr_5a6a8c4c-bbe8-4808-a83a-9f00bcf1919f/ovsdb-server/0.log" Dec 01 22:52:05 crc kubenswrapper[4857]: I1201 22:52:05.689302 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-h6nrk_6640e334-5d6f-441b-b213-985398e67768/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:52:05 crc kubenswrapper[4857]: I1201 22:52:05.700601 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_467e700f-9a49-4c3d-8108-9bb78cc31495/openstack-network-exporter/0.log" Dec 01 22:52:05 crc kubenswrapper[4857]: I1201 22:52:05.811174 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_467e700f-9a49-4c3d-8108-9bb78cc31495/ovn-northd/0.log" Dec 01 22:52:05 crc kubenswrapper[4857]: I1201 22:52:05.923833 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b726f947-1deb-4268-ae76-900394324ecf/openstack-network-exporter/0.log" Dec 01 22:52:06 crc kubenswrapper[4857]: I1201 22:52:06.220118 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b726f947-1deb-4268-ae76-900394324ecf/ovsdbserver-nb/0.log" Dec 01 22:52:06 crc kubenswrapper[4857]: I1201 22:52:06.314271 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3/openstack-network-exporter/0.log" Dec 01 22:52:06 crc kubenswrapper[4857]: I1201 22:52:06.389231 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ecd20b97-c60f-48c4-b2ab-a11ff8fa38f3/ovsdbserver-sb/0.log" Dec 01 22:52:06 crc kubenswrapper[4857]: I1201 22:52:06.598590 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6965455568-h4s4z_338abaa2-77aa-403c-b26b-6a5d8c02ff9f/placement-log/0.log" Dec 01 22:52:06 crc kubenswrapper[4857]: I1201 22:52:06.632619 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6965455568-h4s4z_338abaa2-77aa-403c-b26b-6a5d8c02ff9f/placement-api/0.log" Dec 01 22:52:06 crc kubenswrapper[4857]: I1201 22:52:06.699141 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0203f043-2b36-4611-9b70-1fd54f97a1e3/init-config-reloader/0.log" Dec 01 22:52:06 crc kubenswrapper[4857]: I1201 22:52:06.896449 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0203f043-2b36-4611-9b70-1fd54f97a1e3/config-reloader/0.log" Dec 01 22:52:06 crc kubenswrapper[4857]: I1201 22:52:06.908113 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0203f043-2b36-4611-9b70-1fd54f97a1e3/prometheus/0.log" Dec 01 22:52:06 crc kubenswrapper[4857]: I1201 22:52:06.909926 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0203f043-2b36-4611-9b70-1fd54f97a1e3/init-config-reloader/0.log" Dec 01 22:52:07 crc kubenswrapper[4857]: I1201 22:52:07.009096 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0203f043-2b36-4611-9b70-1fd54f97a1e3/thanos-sidecar/0.log" Dec 01 22:52:07 crc kubenswrapper[4857]: I1201 22:52:07.101894 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825/setup-container/0.log" Dec 01 22:52:07 crc kubenswrapper[4857]: I1201 22:52:07.291722 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825/rabbitmq/0.log" Dec 01 22:52:07 crc kubenswrapper[4857]: I1201 22:52:07.379417 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cbfa87b5-fc91-47a6-8b3a-08a0ba3c9825/setup-container/0.log" Dec 01 22:52:07 crc kubenswrapper[4857]: I1201 22:52:07.409108 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a49ab0a8-09f6-4bfa-9276-c89606c0030e/setup-container/0.log" Dec 01 22:52:07 crc kubenswrapper[4857]: I1201 22:52:07.636791 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a49ab0a8-09f6-4bfa-9276-c89606c0030e/rabbitmq/0.log" Dec 01 22:52:07 crc kubenswrapper[4857]: I1201 22:52:07.649977 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a49ab0a8-09f6-4bfa-9276-c89606c0030e/setup-container/0.log" Dec 01 22:52:07 crc kubenswrapper[4857]: I1201 22:52:07.666837 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-n6dv4_ad6a5aa6-3dc8-4387-9834-7681a0bf77ce/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:52:07 crc kubenswrapper[4857]: I1201 22:52:07.815122 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-twm8k_8ae0c086-76b5-44dc-ab9f-e8988b92cc02/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:52:07 crc kubenswrapper[4857]: I1201 22:52:07.925113 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-zdw5h_bc94c543-00bf-4ce3-a8cc-1ed51f8bbf3c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:52:08 crc kubenswrapper[4857]: I1201 22:52:08.115889 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-cwc7h_a1548cec-634a-469e-b3bf-1813dfe676c1/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:52:08 crc kubenswrapper[4857]: I1201 22:52:08.125614 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8zsck_e7a9f9f9-0f72-46cc-9f30-461fe1a33903/ssh-known-hosts-edpm-deployment/0.log" Dec 01 22:52:08 crc kubenswrapper[4857]: I1201 22:52:08.420688 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b8476f857-zrr4q_b43470bb-ee08-4170-94c6-4a3acbbce50e/proxy-server/0.log" Dec 01 22:52:08 crc kubenswrapper[4857]: I1201 22:52:08.437997 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b8476f857-zrr4q_b43470bb-ee08-4170-94c6-4a3acbbce50e/proxy-httpd/0.log" Dec 01 22:52:08 crc kubenswrapper[4857]: I1201 22:52:08.589823 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-q5szn_34c1c27c-1d55-4a30-ae6a-4e9d5f46b91b/swift-ring-rebalance/0.log" Dec 01 22:52:08 crc kubenswrapper[4857]: I1201 22:52:08.608903 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/account-auditor/0.log" Dec 01 22:52:08 crc kubenswrapper[4857]: I1201 22:52:08.644493 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/account-reaper/0.log" Dec 01 22:52:08 crc kubenswrapper[4857]: I1201 22:52:08.822785 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/account-server/0.log" Dec 01 22:52:08 crc kubenswrapper[4857]: I1201 22:52:08.888979 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/container-auditor/0.log" Dec 01 22:52:08 crc kubenswrapper[4857]: I1201 22:52:08.944847 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/account-replicator/0.log" Dec 01 22:52:08 crc kubenswrapper[4857]: I1201 22:52:08.982293 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/container-replicator/0.log" Dec 01 22:52:09 crc kubenswrapper[4857]: I1201 22:52:09.047390 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/container-server/0.log" Dec 01 22:52:09 crc kubenswrapper[4857]: I1201 22:52:09.124026 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/container-updater/0.log" Dec 01 22:52:09 crc kubenswrapper[4857]: I1201 22:52:09.151799 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/object-auditor/0.log" Dec 01 22:52:09 crc kubenswrapper[4857]: I1201 22:52:09.211315 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/object-expirer/0.log" Dec 01 22:52:09 crc kubenswrapper[4857]: I1201 22:52:09.262890 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/object-replicator/0.log" Dec 01 22:52:09 crc kubenswrapper[4857]: I1201 22:52:09.369167 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/object-server/0.log" Dec 01 22:52:09 crc kubenswrapper[4857]: I1201 22:52:09.414997 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/object-updater/0.log" Dec 01 22:52:09 crc kubenswrapper[4857]: I1201 22:52:09.444842 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/rsync/0.log" Dec 01 22:52:09 crc kubenswrapper[4857]: I1201 22:52:09.471099 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3ea80def-31e0-4cc2-9f15-26ee35c62641/swift-recon-cron/0.log" Dec 01 22:52:09 crc kubenswrapper[4857]: I1201 22:52:09.681442 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-6wnms_2aa8b0d6-1236-48c6-a69c-5996c4a0d6d6/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:52:09 crc kubenswrapper[4857]: I1201 22:52:09.683935 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-sxgs9_e74258df-2d15-4928-993f-fac533426f97/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.417542 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-crwr4"] Dec 01 22:52:11 crc kubenswrapper[4857]: E1201 22:52:11.417971 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd4cd826-888f-40c3-a91c-8982febb154d" containerName="container-00" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.417987 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd4cd826-888f-40c3-a91c-8982febb154d" containerName="container-00" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.418224 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd4cd826-888f-40c3-a91c-8982febb154d" containerName="container-00" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.422527 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.431454 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-crwr4"] Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.590368 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7795b96-a914-4979-a013-dd36e871d693-catalog-content\") pod \"redhat-marketplace-crwr4\" (UID: \"f7795b96-a914-4979-a013-dd36e871d693\") " pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.590440 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxbmj\" (UniqueName: \"kubernetes.io/projected/f7795b96-a914-4979-a013-dd36e871d693-kube-api-access-lxbmj\") pod \"redhat-marketplace-crwr4\" (UID: \"f7795b96-a914-4979-a013-dd36e871d693\") " pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.590471 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7795b96-a914-4979-a013-dd36e871d693-utilities\") pod \"redhat-marketplace-crwr4\" (UID: \"f7795b96-a914-4979-a013-dd36e871d693\") " pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.691873 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7795b96-a914-4979-a013-dd36e871d693-catalog-content\") pod \"redhat-marketplace-crwr4\" (UID: \"f7795b96-a914-4979-a013-dd36e871d693\") " pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.691942 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxbmj\" (UniqueName: \"kubernetes.io/projected/f7795b96-a914-4979-a013-dd36e871d693-kube-api-access-lxbmj\") pod \"redhat-marketplace-crwr4\" (UID: \"f7795b96-a914-4979-a013-dd36e871d693\") " pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.691972 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7795b96-a914-4979-a013-dd36e871d693-utilities\") pod \"redhat-marketplace-crwr4\" (UID: \"f7795b96-a914-4979-a013-dd36e871d693\") " pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.693063 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7795b96-a914-4979-a013-dd36e871d693-catalog-content\") pod \"redhat-marketplace-crwr4\" (UID: \"f7795b96-a914-4979-a013-dd36e871d693\") " pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.693319 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7795b96-a914-4979-a013-dd36e871d693-utilities\") pod \"redhat-marketplace-crwr4\" (UID: \"f7795b96-a914-4979-a013-dd36e871d693\") " pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.711009 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxbmj\" (UniqueName: \"kubernetes.io/projected/f7795b96-a914-4979-a013-dd36e871d693-kube-api-access-lxbmj\") pod \"redhat-marketplace-crwr4\" (UID: \"f7795b96-a914-4979-a013-dd36e871d693\") " pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:11 crc kubenswrapper[4857]: I1201 22:52:11.789511 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:12 crc kubenswrapper[4857]: I1201 22:52:12.278519 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-crwr4"] Dec 01 22:52:13 crc kubenswrapper[4857]: I1201 22:52:13.231812 4857 generic.go:334] "Generic (PLEG): container finished" podID="f7795b96-a914-4979-a013-dd36e871d693" containerID="4bb3861030863cb60ee116a3a680bc45e44785163877f1b7570ef0c225023963" exitCode=0 Dec 01 22:52:13 crc kubenswrapper[4857]: I1201 22:52:13.231915 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crwr4" event={"ID":"f7795b96-a914-4979-a013-dd36e871d693","Type":"ContainerDied","Data":"4bb3861030863cb60ee116a3a680bc45e44785163877f1b7570ef0c225023963"} Dec 01 22:52:13 crc kubenswrapper[4857]: I1201 22:52:13.232075 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crwr4" event={"ID":"f7795b96-a914-4979-a013-dd36e871d693","Type":"ContainerStarted","Data":"0bd3b007f1bc5a71db429e6da42e038b770268447a325d8edfee1436d935099e"} Dec 01 22:52:14 crc kubenswrapper[4857]: I1201 22:52:14.241701 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crwr4" event={"ID":"f7795b96-a914-4979-a013-dd36e871d693","Type":"ContainerStarted","Data":"810c4c7ec25c6d7dcb778572d9c0104dfcad2b1ee7abed2bc4cea78ddc6a6372"} Dec 01 22:52:15 crc kubenswrapper[4857]: I1201 22:52:15.253264 4857 generic.go:334] "Generic (PLEG): container finished" podID="f7795b96-a914-4979-a013-dd36e871d693" containerID="810c4c7ec25c6d7dcb778572d9c0104dfcad2b1ee7abed2bc4cea78ddc6a6372" exitCode=0 Dec 01 22:52:15 crc kubenswrapper[4857]: I1201 22:52:15.253344 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crwr4" event={"ID":"f7795b96-a914-4979-a013-dd36e871d693","Type":"ContainerDied","Data":"810c4c7ec25c6d7dcb778572d9c0104dfcad2b1ee7abed2bc4cea78ddc6a6372"} Dec 01 22:52:16 crc kubenswrapper[4857]: I1201 22:52:16.290660 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crwr4" event={"ID":"f7795b96-a914-4979-a013-dd36e871d693","Type":"ContainerStarted","Data":"85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5"} Dec 01 22:52:16 crc kubenswrapper[4857]: I1201 22:52:16.311087 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-crwr4" podStartSLOduration=2.728126656 podStartE2EDuration="5.311061964s" podCreationTimestamp="2025-12-01 22:52:11 +0000 UTC" firstStartedPulling="2025-12-01 22:52:13.233491974 +0000 UTC m=+4651.723554291" lastFinishedPulling="2025-12-01 22:52:15.816427282 +0000 UTC m=+4654.306489599" observedRunningTime="2025-12-01 22:52:16.310007058 +0000 UTC m=+4654.800069365" watchObservedRunningTime="2025-12-01 22:52:16.311061964 +0000 UTC m=+4654.801124281" Dec 01 22:52:16 crc kubenswrapper[4857]: I1201 22:52:16.834729 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:52:16 crc kubenswrapper[4857]: E1201 22:52:16.835003 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:52:17 crc kubenswrapper[4857]: I1201 22:52:17.947892 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_6a290cd2-c321-4cd5-b168-4866c03b135b/memcached/0.log" Dec 01 22:52:21 crc kubenswrapper[4857]: I1201 22:52:21.790570 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:21 crc kubenswrapper[4857]: I1201 22:52:21.791022 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:21 crc kubenswrapper[4857]: I1201 22:52:21.850913 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:22 crc kubenswrapper[4857]: I1201 22:52:22.945194 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:22 crc kubenswrapper[4857]: I1201 22:52:22.994244 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-crwr4"] Dec 01 22:52:24 crc kubenswrapper[4857]: I1201 22:52:24.353631 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-crwr4" podUID="f7795b96-a914-4979-a013-dd36e871d693" containerName="registry-server" containerID="cri-o://85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5" gracePeriod=2 Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.028911 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.203199 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxbmj\" (UniqueName: \"kubernetes.io/projected/f7795b96-a914-4979-a013-dd36e871d693-kube-api-access-lxbmj\") pod \"f7795b96-a914-4979-a013-dd36e871d693\" (UID: \"f7795b96-a914-4979-a013-dd36e871d693\") " Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.203246 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7795b96-a914-4979-a013-dd36e871d693-utilities\") pod \"f7795b96-a914-4979-a013-dd36e871d693\" (UID: \"f7795b96-a914-4979-a013-dd36e871d693\") " Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.203479 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7795b96-a914-4979-a013-dd36e871d693-catalog-content\") pod \"f7795b96-a914-4979-a013-dd36e871d693\" (UID: \"f7795b96-a914-4979-a013-dd36e871d693\") " Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.204171 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7795b96-a914-4979-a013-dd36e871d693-utilities" (OuterVolumeSpecName: "utilities") pod "f7795b96-a914-4979-a013-dd36e871d693" (UID: "f7795b96-a914-4979-a013-dd36e871d693"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.210200 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7795b96-a914-4979-a013-dd36e871d693-kube-api-access-lxbmj" (OuterVolumeSpecName: "kube-api-access-lxbmj") pod "f7795b96-a914-4979-a013-dd36e871d693" (UID: "f7795b96-a914-4979-a013-dd36e871d693"). InnerVolumeSpecName "kube-api-access-lxbmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.223754 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7795b96-a914-4979-a013-dd36e871d693-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7795b96-a914-4979-a013-dd36e871d693" (UID: "f7795b96-a914-4979-a013-dd36e871d693"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.305845 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7795b96-a914-4979-a013-dd36e871d693-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.305873 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxbmj\" (UniqueName: \"kubernetes.io/projected/f7795b96-a914-4979-a013-dd36e871d693-kube-api-access-lxbmj\") on node \"crc\" DevicePath \"\"" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.305898 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7795b96-a914-4979-a013-dd36e871d693-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.364780 4857 generic.go:334] "Generic (PLEG): container finished" podID="f7795b96-a914-4979-a013-dd36e871d693" containerID="85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5" exitCode=0 Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.364823 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crwr4" event={"ID":"f7795b96-a914-4979-a013-dd36e871d693","Type":"ContainerDied","Data":"85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5"} Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.364850 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crwr4" event={"ID":"f7795b96-a914-4979-a013-dd36e871d693","Type":"ContainerDied","Data":"0bd3b007f1bc5a71db429e6da42e038b770268447a325d8edfee1436d935099e"} Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.364868 4857 scope.go:117] "RemoveContainer" containerID="85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.365064 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-crwr4" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.385785 4857 scope.go:117] "RemoveContainer" containerID="810c4c7ec25c6d7dcb778572d9c0104dfcad2b1ee7abed2bc4cea78ddc6a6372" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.415562 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-crwr4"] Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.435759 4857 scope.go:117] "RemoveContainer" containerID="4bb3861030863cb60ee116a3a680bc45e44785163877f1b7570ef0c225023963" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.436402 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-crwr4"] Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.460573 4857 scope.go:117] "RemoveContainer" containerID="85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5" Dec 01 22:52:25 crc kubenswrapper[4857]: E1201 22:52:25.461183 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5\": container with ID starting with 85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5 not found: ID does not exist" containerID="85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.461230 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5"} err="failed to get container status \"85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5\": rpc error: code = NotFound desc = could not find container \"85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5\": container with ID starting with 85742d311131a36b57df63e6104f8dbc3436e3ca7f4d8d775f0dbdc296e198f5 not found: ID does not exist" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.461256 4857 scope.go:117] "RemoveContainer" containerID="810c4c7ec25c6d7dcb778572d9c0104dfcad2b1ee7abed2bc4cea78ddc6a6372" Dec 01 22:52:25 crc kubenswrapper[4857]: E1201 22:52:25.461658 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"810c4c7ec25c6d7dcb778572d9c0104dfcad2b1ee7abed2bc4cea78ddc6a6372\": container with ID starting with 810c4c7ec25c6d7dcb778572d9c0104dfcad2b1ee7abed2bc4cea78ddc6a6372 not found: ID does not exist" containerID="810c4c7ec25c6d7dcb778572d9c0104dfcad2b1ee7abed2bc4cea78ddc6a6372" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.461691 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"810c4c7ec25c6d7dcb778572d9c0104dfcad2b1ee7abed2bc4cea78ddc6a6372"} err="failed to get container status \"810c4c7ec25c6d7dcb778572d9c0104dfcad2b1ee7abed2bc4cea78ddc6a6372\": rpc error: code = NotFound desc = could not find container \"810c4c7ec25c6d7dcb778572d9c0104dfcad2b1ee7abed2bc4cea78ddc6a6372\": container with ID starting with 810c4c7ec25c6d7dcb778572d9c0104dfcad2b1ee7abed2bc4cea78ddc6a6372 not found: ID does not exist" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.461713 4857 scope.go:117] "RemoveContainer" containerID="4bb3861030863cb60ee116a3a680bc45e44785163877f1b7570ef0c225023963" Dec 01 22:52:25 crc kubenswrapper[4857]: E1201 22:52:25.461934 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bb3861030863cb60ee116a3a680bc45e44785163877f1b7570ef0c225023963\": container with ID starting with 4bb3861030863cb60ee116a3a680bc45e44785163877f1b7570ef0c225023963 not found: ID does not exist" containerID="4bb3861030863cb60ee116a3a680bc45e44785163877f1b7570ef0c225023963" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.461952 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bb3861030863cb60ee116a3a680bc45e44785163877f1b7570ef0c225023963"} err="failed to get container status \"4bb3861030863cb60ee116a3a680bc45e44785163877f1b7570ef0c225023963\": rpc error: code = NotFound desc = could not find container \"4bb3861030863cb60ee116a3a680bc45e44785163877f1b7570ef0c225023963\": container with ID starting with 4bb3861030863cb60ee116a3a680bc45e44785163877f1b7570ef0c225023963 not found: ID does not exist" Dec 01 22:52:25 crc kubenswrapper[4857]: I1201 22:52:25.845364 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7795b96-a914-4979-a013-dd36e871d693" path="/var/lib/kubelet/pods/f7795b96-a914-4979-a013-dd36e871d693/volumes" Dec 01 22:52:31 crc kubenswrapper[4857]: I1201 22:52:31.848653 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:52:31 crc kubenswrapper[4857]: E1201 22:52:31.849501 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:52:41 crc kubenswrapper[4857]: I1201 22:52:41.944735 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/util/0.log" Dec 01 22:52:42 crc kubenswrapper[4857]: I1201 22:52:42.141010 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/util/0.log" Dec 01 22:52:42 crc kubenswrapper[4857]: I1201 22:52:42.161582 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/pull/0.log" Dec 01 22:52:42 crc kubenswrapper[4857]: I1201 22:52:42.202489 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/pull/0.log" Dec 01 22:52:42 crc kubenswrapper[4857]: I1201 22:52:42.376337 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/pull/0.log" Dec 01 22:52:42 crc kubenswrapper[4857]: I1201 22:52:42.396670 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/extract/0.log" Dec 01 22:52:42 crc kubenswrapper[4857]: I1201 22:52:42.413645 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_20d2b11f39a764cfc6eafc23757971bc68cc4422fce033f5d1fb9ea198vg2ms_8ff7e678-7e85-4ba7-a5cb-d332cb47564a/util/0.log" Dec 01 22:52:42 crc kubenswrapper[4857]: I1201 22:52:42.556751 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-rqfzk_80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4/kube-rbac-proxy/0.log" Dec 01 22:52:42 crc kubenswrapper[4857]: I1201 22:52:42.644882 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-rqfzk_80e71f2f-8dc8-4ee0-a86d-e8e5756f0ae4/manager/0.log" Dec 01 22:52:42 crc kubenswrapper[4857]: I1201 22:52:42.655678 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-5gqcm_d3a93b19-a249-499a-89fc-9b2a1a69c1e0/kube-rbac-proxy/0.log" Dec 01 22:52:42 crc kubenswrapper[4857]: I1201 22:52:42.801657 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-5gqcm_d3a93b19-a249-499a-89fc-9b2a1a69c1e0/manager/0.log" Dec 01 22:52:42 crc kubenswrapper[4857]: I1201 22:52:42.878254 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-dz5k5_393fd1a5-e3b9-44d6-a9ef-1447a0f9f465/kube-rbac-proxy/0.log" Dec 01 22:52:42 crc kubenswrapper[4857]: I1201 22:52:42.879753 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-dz5k5_393fd1a5-e3b9-44d6-a9ef-1447a0f9f465/manager/0.log" Dec 01 22:52:43 crc kubenswrapper[4857]: I1201 22:52:43.072803 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-9qwbq_df91f733-711d-43b0-98a0-b3fa6b119fbe/kube-rbac-proxy/0.log" Dec 01 22:52:43 crc kubenswrapper[4857]: I1201 22:52:43.104799 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-9qwbq_df91f733-711d-43b0-98a0-b3fa6b119fbe/manager/0.log" Dec 01 22:52:43 crc kubenswrapper[4857]: I1201 22:52:43.189229 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-zkqw9_068a21df-001f-45e6-9946-68f67bad7d43/kube-rbac-proxy/0.log" Dec 01 22:52:43 crc kubenswrapper[4857]: I1201 22:52:43.284257 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-zkqw9_068a21df-001f-45e6-9946-68f67bad7d43/manager/0.log" Dec 01 22:52:43 crc kubenswrapper[4857]: I1201 22:52:43.319944 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-4g68b_b972200d-235d-4ae5-a71d-dd206e4f6025/kube-rbac-proxy/0.log" Dec 01 22:52:43 crc kubenswrapper[4857]: I1201 22:52:43.378880 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-4g68b_b972200d-235d-4ae5-a71d-dd206e4f6025/manager/0.log" Dec 01 22:52:43 crc kubenswrapper[4857]: I1201 22:52:43.607204 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-54xtc_23e7348c-65d9-4c4b-9c1e-d86cc3b55bff/kube-rbac-proxy/0.log" Dec 01 22:52:43 crc kubenswrapper[4857]: I1201 22:52:43.782671 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-54xtc_23e7348c-65d9-4c4b-9c1e-d86cc3b55bff/manager/0.log" Dec 01 22:52:43 crc kubenswrapper[4857]: I1201 22:52:43.807587 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-j277f_5dd6bfb5-2edc-49b4-8877-85ab5722cdf0/kube-rbac-proxy/0.log" Dec 01 22:52:43 crc kubenswrapper[4857]: I1201 22:52:43.837351 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-j277f_5dd6bfb5-2edc-49b4-8877-85ab5722cdf0/manager/0.log" Dec 01 22:52:43 crc kubenswrapper[4857]: I1201 22:52:43.998146 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-vtcx7_dcef9388-5d9b-4172-bb80-ec69b3e98dd0/kube-rbac-proxy/0.log" Dec 01 22:52:44 crc kubenswrapper[4857]: I1201 22:52:44.092213 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-vtcx7_dcef9388-5d9b-4172-bb80-ec69b3e98dd0/manager/0.log" Dec 01 22:52:44 crc kubenswrapper[4857]: I1201 22:52:44.097809 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-b7ksb_92c43db5-8128-4269-9ad4-a263de88f450/kube-rbac-proxy/0.log" Dec 01 22:52:44 crc kubenswrapper[4857]: I1201 22:52:44.210845 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-b7ksb_92c43db5-8128-4269-9ad4-a263de88f450/manager/0.log" Dec 01 22:52:44 crc kubenswrapper[4857]: I1201 22:52:44.279146 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-f9d2t_b13a13cc-0689-480a-bc95-e19aba327f9c/kube-rbac-proxy/0.log" Dec 01 22:52:44 crc kubenswrapper[4857]: I1201 22:52:44.291277 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-f9d2t_b13a13cc-0689-480a-bc95-e19aba327f9c/manager/0.log" Dec 01 22:52:44 crc kubenswrapper[4857]: I1201 22:52:44.567079 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-p6v4v_188a7d67-ffec-4608-9b5f-cc33972e6fed/kube-rbac-proxy/0.log" Dec 01 22:52:44 crc kubenswrapper[4857]: I1201 22:52:44.689028 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-p6v4v_188a7d67-ffec-4608-9b5f-cc33972e6fed/manager/0.log" Dec 01 22:52:44 crc kubenswrapper[4857]: I1201 22:52:44.759257 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vbvnt_de53c246-6e20-4b67-be68-1c5356fd46d4/kube-rbac-proxy/0.log" Dec 01 22:52:44 crc kubenswrapper[4857]: I1201 22:52:44.853082 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vbvnt_de53c246-6e20-4b67-be68-1c5356fd46d4/manager/0.log" Dec 01 22:52:44 crc kubenswrapper[4857]: I1201 22:52:44.870519 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-cncd6_246a4bf0-515b-46f8-bd7f-89e7e961d774/kube-rbac-proxy/0.log" Dec 01 22:52:44 crc kubenswrapper[4857]: I1201 22:52:44.997508 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-cncd6_246a4bf0-515b-46f8-bd7f-89e7e961d774/manager/0.log" Dec 01 22:52:45 crc kubenswrapper[4857]: I1201 22:52:45.052756 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c_1c2d54cf-e53d-483d-9b00-6f03858806ab/kube-rbac-proxy/0.log" Dec 01 22:52:45 crc kubenswrapper[4857]: I1201 22:52:45.116663 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4j8q6c_1c2d54cf-e53d-483d-9b00-6f03858806ab/manager/0.log" Dec 01 22:52:45 crc kubenswrapper[4857]: I1201 22:52:45.512859 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-9f7fd9cfd-66kxd_1e12f76f-2a0a-46d8-aec6-6d9a7cf7e0dc/operator/0.log" Dec 01 22:52:45 crc kubenswrapper[4857]: I1201 22:52:45.580866 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-m6lxm_f8b95808-7d34-4097-bac8-f7e081d9c039/registry-server/0.log" Dec 01 22:52:45 crc kubenswrapper[4857]: I1201 22:52:45.806466 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-7vmng_f6e43ef6-1269-40f5-af4d-fe2f30dea319/manager/0.log" Dec 01 22:52:45 crc kubenswrapper[4857]: I1201 22:52:45.830762 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-mww86_c5ab847c-10d3-482d-8c70-e3ab439e7b5d/kube-rbac-proxy/0.log" Dec 01 22:52:45 crc kubenswrapper[4857]: I1201 22:52:45.834599 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:52:45 crc kubenswrapper[4857]: E1201 22:52:45.834875 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:52:45 crc kubenswrapper[4857]: I1201 22:52:45.856053 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-7vmng_f6e43ef6-1269-40f5-af4d-fe2f30dea319/kube-rbac-proxy/0.log" Dec 01 22:52:46 crc kubenswrapper[4857]: I1201 22:52:46.051789 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-mww86_c5ab847c-10d3-482d-8c70-e3ab439e7b5d/manager/0.log" Dec 01 22:52:46 crc kubenswrapper[4857]: I1201 22:52:46.173860 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-gdqjh_1d490d1b-558b-49fb-9b0f-65d0f82b0148/operator/0.log" Dec 01 22:52:46 crc kubenswrapper[4857]: I1201 22:52:46.299193 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-wjnzv_cdd278ff-9eb6-45f7-977d-02944f6f312a/kube-rbac-proxy/0.log" Dec 01 22:52:46 crc kubenswrapper[4857]: I1201 22:52:46.327019 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-wjnzv_cdd278ff-9eb6-45f7-977d-02944f6f312a/manager/0.log" Dec 01 22:52:46 crc kubenswrapper[4857]: I1201 22:52:46.422556 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f765fc7d6-x797d_ff230e58-1711-4638-89d6-daf91fb3bfc1/kube-rbac-proxy/0.log" Dec 01 22:52:46 crc kubenswrapper[4857]: I1201 22:52:46.465259 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-777b97d487-s7g6x_0a4c6432-bb6c-41b6-9981-ef6ac1354ae3/manager/0.log" Dec 01 22:52:46 crc kubenswrapper[4857]: I1201 22:52:46.614811 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f765fc7d6-x797d_ff230e58-1711-4638-89d6-daf91fb3bfc1/manager/0.log" Dec 01 22:52:46 crc kubenswrapper[4857]: I1201 22:52:46.616486 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wccvn_4c8d087e-e00c-4dad-87c7-06ab1017abaf/manager/0.log" Dec 01 22:52:46 crc kubenswrapper[4857]: I1201 22:52:46.619196 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wccvn_4c8d087e-e00c-4dad-87c7-06ab1017abaf/kube-rbac-proxy/0.log" Dec 01 22:52:46 crc kubenswrapper[4857]: I1201 22:52:46.760030 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-vh887_8a6276de-e994-4f62-a625-906c9a5f4f5d/kube-rbac-proxy/0.log" Dec 01 22:52:46 crc kubenswrapper[4857]: I1201 22:52:46.785489 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-vh887_8a6276de-e994-4f62-a625-906c9a5f4f5d/manager/0.log" Dec 01 22:52:56 crc kubenswrapper[4857]: I1201 22:52:56.835502 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:52:56 crc kubenswrapper[4857]: E1201 22:52:56.836689 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:53:10 crc kubenswrapper[4857]: I1201 22:53:10.835705 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:53:10 crc kubenswrapper[4857]: E1201 22:53:10.836551 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:53:15 crc kubenswrapper[4857]: I1201 22:53:15.790661 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-8mmlp_1cb66d76-6c0a-48c9-a0a8-54a60f7aae01/control-plane-machine-set-operator/0.log" Dec 01 22:53:16 crc kubenswrapper[4857]: I1201 22:53:16.541159 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-2bwxl_00d4b52b-5372-48a1-a1b4-6c17e18cc7a4/kube-rbac-proxy/0.log" Dec 01 22:53:16 crc kubenswrapper[4857]: I1201 22:53:16.599712 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-2bwxl_00d4b52b-5372-48a1-a1b4-6c17e18cc7a4/machine-api-operator/0.log" Dec 01 22:53:24 crc kubenswrapper[4857]: I1201 22:53:24.836126 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:53:24 crc kubenswrapper[4857]: E1201 22:53:24.837520 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:53:31 crc kubenswrapper[4857]: I1201 22:53:31.819810 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-wnq4k_e8804d0a-50d0-4b87-bc77-a8c8229ccf1d/cert-manager-controller/0.log" Dec 01 22:53:31 crc kubenswrapper[4857]: I1201 22:53:31.975611 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-lcww8_84a659be-d476-4ba1-820e-c5279704aec5/cert-manager-cainjector/0.log" Dec 01 22:53:31 crc kubenswrapper[4857]: I1201 22:53:31.983283 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-5gv57_64fd8572-2a45-422e-a413-49314dc52a55/cert-manager-webhook/0.log" Dec 01 22:53:37 crc kubenswrapper[4857]: I1201 22:53:37.835064 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:53:37 crc kubenswrapper[4857]: E1201 22:53:37.835699 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:53:50 crc kubenswrapper[4857]: I1201 22:53:50.683876 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-78hdv_0764bcb8-a3f2-46cd-bceb-f08b51012b7d/nmstate-console-plugin/0.log" Dec 01 22:53:50 crc kubenswrapper[4857]: I1201 22:53:50.867889 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-grx52_e37b046b-176a-4bc3-bcbc-3093fe08b256/nmstate-handler/0.log" Dec 01 22:53:50 crc kubenswrapper[4857]: I1201 22:53:50.916633 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-m49jm_c7894b0e-91be-4ed2-810b-f9e24df87fae/nmstate-metrics/0.log" Dec 01 22:53:50 crc kubenswrapper[4857]: I1201 22:53:50.930732 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-m49jm_c7894b0e-91be-4ed2-810b-f9e24df87fae/kube-rbac-proxy/0.log" Dec 01 22:53:51 crc kubenswrapper[4857]: I1201 22:53:51.048454 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-m8q4r_8c989c5f-b814-437a-bdca-86e63044ce91/nmstate-operator/0.log" Dec 01 22:53:51 crc kubenswrapper[4857]: I1201 22:53:51.150224 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-bdx6w_79431306-c7c5-4c07-8b52-ec94fe45e6da/nmstate-webhook/0.log" Dec 01 22:53:52 crc kubenswrapper[4857]: I1201 22:53:52.835759 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:53:52 crc kubenswrapper[4857]: E1201 22:53:52.836376 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:54:03 crc kubenswrapper[4857]: I1201 22:54:03.834725 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:54:03 crc kubenswrapper[4857]: E1201 22:54:03.835489 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:54:09 crc kubenswrapper[4857]: I1201 22:54:09.911791 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-2gzlv_28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4/kube-rbac-proxy/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.097389 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-2gzlv_28caf5c7-5c41-4a3b-95a1-75ba12ca3bb4/controller/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.129602 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-x6nqn_ee319d44-c1a3-4325-835f-3d9fc6a187fd/frr-k8s-webhook-server/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.260231 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-frr-files/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.514276 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-metrics/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.517293 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-reloader/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.524390 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-frr-files/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.577968 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-reloader/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.709386 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-reloader/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.718958 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-frr-files/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.773332 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-metrics/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.775646 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-metrics/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.946461 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-frr-files/0.log" Dec 01 22:54:10 crc kubenswrapper[4857]: I1201 22:54:10.950196 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-reloader/0.log" Dec 01 22:54:11 crc kubenswrapper[4857]: I1201 22:54:11.022210 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/controller/0.log" Dec 01 22:54:11 crc kubenswrapper[4857]: I1201 22:54:11.023212 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/cp-metrics/0.log" Dec 01 22:54:11 crc kubenswrapper[4857]: I1201 22:54:11.184646 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/frr-metrics/0.log" Dec 01 22:54:11 crc kubenswrapper[4857]: I1201 22:54:11.205450 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/kube-rbac-proxy/0.log" Dec 01 22:54:11 crc kubenswrapper[4857]: I1201 22:54:11.242126 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/kube-rbac-proxy-frr/0.log" Dec 01 22:54:11 crc kubenswrapper[4857]: I1201 22:54:11.383501 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/reloader/0.log" Dec 01 22:54:11 crc kubenswrapper[4857]: I1201 22:54:11.457219 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-75c9cd8c7-tt4cx_8eee3586-4292-4165-81c4-496c2ee8ab9c/manager/0.log" Dec 01 22:54:11 crc kubenswrapper[4857]: I1201 22:54:11.624450 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-75b5f488d8-k8rxs_bd5afa07-4d43-4dce-8292-58ad63e6362a/webhook-server/0.log" Dec 01 22:54:11 crc kubenswrapper[4857]: I1201 22:54:11.803545 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r64qq_c6620f92-fe9f-4dde-a758-b46475bd8db8/kube-rbac-proxy/0.log" Dec 01 22:54:12 crc kubenswrapper[4857]: I1201 22:54:12.442689 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r64qq_c6620f92-fe9f-4dde-a758-b46475bd8db8/speaker/0.log" Dec 01 22:54:12 crc kubenswrapper[4857]: I1201 22:54:12.846688 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjx7t_994ed59f-d2c6-4af0-b868-f167195e071c/frr/0.log" Dec 01 22:54:16 crc kubenswrapper[4857]: I1201 22:54:16.835997 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:54:16 crc kubenswrapper[4857]: E1201 22:54:16.837192 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:54:27 crc kubenswrapper[4857]: I1201 22:54:27.809538 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/util/0.log" Dec 01 22:54:27 crc kubenswrapper[4857]: I1201 22:54:27.835560 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:54:27 crc kubenswrapper[4857]: E1201 22:54:27.835928 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.010109 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/util/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.069257 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/pull/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.103271 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/pull/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.235193 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/pull/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.239083 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/util/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.246113 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fwpszk_d0bf962d-59bc-40cf-b395-e5cd376eda23/extract/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.400316 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/util/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.584125 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/pull/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.600616 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/pull/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.624373 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/util/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.780854 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/pull/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.804231 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/util/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.812476 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fsjhd_a058ee2b-6a45-4dbe-b0dd-1c6c5a0d8ba8/extract/0.log" Dec 01 22:54:28 crc kubenswrapper[4857]: I1201 22:54:28.954187 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/util/0.log" Dec 01 22:54:29 crc kubenswrapper[4857]: I1201 22:54:29.195545 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/util/0.log" Dec 01 22:54:29 crc kubenswrapper[4857]: I1201 22:54:29.212452 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/pull/0.log" Dec 01 22:54:29 crc kubenswrapper[4857]: I1201 22:54:29.212589 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/pull/0.log" Dec 01 22:54:29 crc kubenswrapper[4857]: I1201 22:54:29.405441 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/util/0.log" Dec 01 22:54:29 crc kubenswrapper[4857]: I1201 22:54:29.406259 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/pull/0.log" Dec 01 22:54:29 crc kubenswrapper[4857]: I1201 22:54:29.422927 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83twc4q_b421a90d-3d2c-48a5-b905-ed356dd54da2/extract/0.log" Dec 01 22:54:29 crc kubenswrapper[4857]: I1201 22:54:29.551692 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/extract-utilities/0.log" Dec 01 22:54:29 crc kubenswrapper[4857]: I1201 22:54:29.731344 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/extract-content/0.log" Dec 01 22:54:29 crc kubenswrapper[4857]: I1201 22:54:29.733209 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/extract-content/0.log" Dec 01 22:54:29 crc kubenswrapper[4857]: I1201 22:54:29.734966 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/extract-utilities/0.log" Dec 01 22:54:29 crc kubenswrapper[4857]: I1201 22:54:29.933178 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/extract-utilities/0.log" Dec 01 22:54:29 crc kubenswrapper[4857]: I1201 22:54:29.942736 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/extract-content/0.log" Dec 01 22:54:30 crc kubenswrapper[4857]: I1201 22:54:30.107126 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/extract-utilities/0.log" Dec 01 22:54:30 crc kubenswrapper[4857]: I1201 22:54:30.486140 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8qfd_c3e413ef-6fbc-4e78-8650-357ea026b79c/registry-server/0.log" Dec 01 22:54:30 crc kubenswrapper[4857]: I1201 22:54:30.748261 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/extract-utilities/0.log" Dec 01 22:54:30 crc kubenswrapper[4857]: I1201 22:54:30.758327 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/extract-content/0.log" Dec 01 22:54:30 crc kubenswrapper[4857]: I1201 22:54:30.794088 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/extract-content/0.log" Dec 01 22:54:30 crc kubenswrapper[4857]: I1201 22:54:30.905022 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/extract-content/0.log" Dec 01 22:54:30 crc kubenswrapper[4857]: I1201 22:54:30.925134 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/extract-utilities/0.log" Dec 01 22:54:31 crc kubenswrapper[4857]: I1201 22:54:31.121851 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/extract-utilities/0.log" Dec 01 22:54:31 crc kubenswrapper[4857]: I1201 22:54:31.132115 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-d5gjm_9fe4a11e-f252-4f61-a8f4-d513b2b40643/marketplace-operator/0.log" Dec 01 22:54:31 crc kubenswrapper[4857]: I1201 22:54:31.375207 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/extract-content/0.log" Dec 01 22:54:31 crc kubenswrapper[4857]: I1201 22:54:31.398049 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/extract-content/0.log" Dec 01 22:54:31 crc kubenswrapper[4857]: I1201 22:54:31.436748 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/extract-utilities/0.log" Dec 01 22:54:31 crc kubenswrapper[4857]: I1201 22:54:31.578798 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mw8ff_00a902c7-b8d4-4f4d-bc21-875579a6f0d0/registry-server/0.log" Dec 01 22:54:31 crc kubenswrapper[4857]: I1201 22:54:31.620383 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/extract-content/0.log" Dec 01 22:54:31 crc kubenswrapper[4857]: I1201 22:54:31.642957 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/extract-utilities/0.log" Dec 01 22:54:31 crc kubenswrapper[4857]: I1201 22:54:31.873157 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2x8zg_c20d9041-cb65-4743-ac7b-3f0f4675b71e/registry-server/0.log" Dec 01 22:54:31 crc kubenswrapper[4857]: I1201 22:54:31.899097 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/extract-utilities/0.log" Dec 01 22:54:32 crc kubenswrapper[4857]: I1201 22:54:32.574483 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/extract-content/0.log" Dec 01 22:54:32 crc kubenswrapper[4857]: I1201 22:54:32.608807 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/extract-content/0.log" Dec 01 22:54:32 crc kubenswrapper[4857]: I1201 22:54:32.614271 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/extract-utilities/0.log" Dec 01 22:54:32 crc kubenswrapper[4857]: I1201 22:54:32.773865 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/extract-content/0.log" Dec 01 22:54:32 crc kubenswrapper[4857]: I1201 22:54:32.795195 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/extract-utilities/0.log" Dec 01 22:54:33 crc kubenswrapper[4857]: I1201 22:54:33.336928 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfmgr_0b7910c3-1306-4e26-a5d6-95a78e751adc/registry-server/0.log" Dec 01 22:54:42 crc kubenswrapper[4857]: I1201 22:54:42.835181 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:54:42 crc kubenswrapper[4857]: E1201 22:54:42.836068 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:54:46 crc kubenswrapper[4857]: I1201 22:54:46.907561 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-tqws8_45d679d3-a9f8-402a-a8cf-4f9913ea40a1/prometheus-operator/0.log" Dec 01 22:54:47 crc kubenswrapper[4857]: I1201 22:54:47.026717 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-744b77886-6c586_203588e4-df3f-407e-9454-9cff35f281e5/prometheus-operator-admission-webhook/0.log" Dec 01 22:54:47 crc kubenswrapper[4857]: I1201 22:54:47.060763 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-744b77886-qczzg_e212d47c-24b3-44cb-a1de-3a291777358d/prometheus-operator-admission-webhook/0.log" Dec 01 22:54:47 crc kubenswrapper[4857]: I1201 22:54:47.285380 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-bxmst_f57d34d6-6aa9-4c9c-a653-cfcda5e7f98f/operator/0.log" Dec 01 22:54:47 crc kubenswrapper[4857]: I1201 22:54:47.290222 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-26djf_f5a63bac-285c-401a-a975-3079cbad38cf/perses-operator/0.log" Dec 01 22:54:54 crc kubenswrapper[4857]: I1201 22:54:54.835328 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:54:54 crc kubenswrapper[4857]: E1201 22:54:54.836967 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.233145 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nl5jx"] Dec 01 22:55:06 crc kubenswrapper[4857]: E1201 22:55:06.234199 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7795b96-a914-4979-a013-dd36e871d693" containerName="registry-server" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.234223 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7795b96-a914-4979-a013-dd36e871d693" containerName="registry-server" Dec 01 22:55:06 crc kubenswrapper[4857]: E1201 22:55:06.234236 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7795b96-a914-4979-a013-dd36e871d693" containerName="extract-utilities" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.234244 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7795b96-a914-4979-a013-dd36e871d693" containerName="extract-utilities" Dec 01 22:55:06 crc kubenswrapper[4857]: E1201 22:55:06.244894 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7795b96-a914-4979-a013-dd36e871d693" containerName="extract-content" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.244919 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7795b96-a914-4979-a013-dd36e871d693" containerName="extract-content" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.246494 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7795b96-a914-4979-a013-dd36e871d693" containerName="registry-server" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.248495 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nl5jx"] Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.248613 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.326410 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d27e720-9a0c-419b-958c-a95d40c04774-catalog-content\") pod \"certified-operators-nl5jx\" (UID: \"3d27e720-9a0c-419b-958c-a95d40c04774\") " pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.326775 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d27e720-9a0c-419b-958c-a95d40c04774-utilities\") pod \"certified-operators-nl5jx\" (UID: \"3d27e720-9a0c-419b-958c-a95d40c04774\") " pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.326840 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmnxb\" (UniqueName: \"kubernetes.io/projected/3d27e720-9a0c-419b-958c-a95d40c04774-kube-api-access-dmnxb\") pod \"certified-operators-nl5jx\" (UID: \"3d27e720-9a0c-419b-958c-a95d40c04774\") " pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.402463 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hj2km"] Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.404432 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.424042 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hj2km"] Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.428491 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d27e720-9a0c-419b-958c-a95d40c04774-catalog-content\") pod \"certified-operators-nl5jx\" (UID: \"3d27e720-9a0c-419b-958c-a95d40c04774\") " pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.428696 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d27e720-9a0c-419b-958c-a95d40c04774-utilities\") pod \"certified-operators-nl5jx\" (UID: \"3d27e720-9a0c-419b-958c-a95d40c04774\") " pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.429143 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrmqd\" (UniqueName: \"kubernetes.io/projected/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-kube-api-access-nrmqd\") pod \"community-operators-hj2km\" (UID: \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\") " pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.429290 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmnxb\" (UniqueName: \"kubernetes.io/projected/3d27e720-9a0c-419b-958c-a95d40c04774-kube-api-access-dmnxb\") pod \"certified-operators-nl5jx\" (UID: \"3d27e720-9a0c-419b-958c-a95d40c04774\") " pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.429462 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d27e720-9a0c-419b-958c-a95d40c04774-utilities\") pod \"certified-operators-nl5jx\" (UID: \"3d27e720-9a0c-419b-958c-a95d40c04774\") " pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.429499 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-catalog-content\") pod \"community-operators-hj2km\" (UID: \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\") " pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.429674 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-utilities\") pod \"community-operators-hj2km\" (UID: \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\") " pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.429506 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d27e720-9a0c-419b-958c-a95d40c04774-catalog-content\") pod \"certified-operators-nl5jx\" (UID: \"3d27e720-9a0c-419b-958c-a95d40c04774\") " pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.531714 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrmqd\" (UniqueName: \"kubernetes.io/projected/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-kube-api-access-nrmqd\") pod \"community-operators-hj2km\" (UID: \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\") " pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.532247 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-catalog-content\") pod \"community-operators-hj2km\" (UID: \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\") " pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.532391 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-utilities\") pod \"community-operators-hj2km\" (UID: \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\") " pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.700536 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-catalog-content\") pod \"community-operators-hj2km\" (UID: \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\") " pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.702108 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-utilities\") pod \"community-operators-hj2km\" (UID: \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\") " pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.957392 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmnxb\" (UniqueName: \"kubernetes.io/projected/3d27e720-9a0c-419b-958c-a95d40c04774-kube-api-access-dmnxb\") pod \"certified-operators-nl5jx\" (UID: \"3d27e720-9a0c-419b-958c-a95d40c04774\") " pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:06 crc kubenswrapper[4857]: I1201 22:55:06.957725 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrmqd\" (UniqueName: \"kubernetes.io/projected/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-kube-api-access-nrmqd\") pod \"community-operators-hj2km\" (UID: \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\") " pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:07 crc kubenswrapper[4857]: I1201 22:55:07.022953 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:07 crc kubenswrapper[4857]: I1201 22:55:07.192451 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:07 crc kubenswrapper[4857]: I1201 22:55:07.628847 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hj2km"] Dec 01 22:55:07 crc kubenswrapper[4857]: I1201 22:55:07.781715 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nl5jx"] Dec 01 22:55:08 crc kubenswrapper[4857]: I1201 22:55:08.191674 4857 generic.go:334] "Generic (PLEG): container finished" podID="e2d1a4d5-3a08-4d66-8207-ef7e243b2121" containerID="385c55244b709a126d5e0176d5747f9754a818b344030371398926a18118f3dc" exitCode=0 Dec 01 22:55:08 crc kubenswrapper[4857]: I1201 22:55:08.191745 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hj2km" event={"ID":"e2d1a4d5-3a08-4d66-8207-ef7e243b2121","Type":"ContainerDied","Data":"385c55244b709a126d5e0176d5747f9754a818b344030371398926a18118f3dc"} Dec 01 22:55:08 crc kubenswrapper[4857]: I1201 22:55:08.191779 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hj2km" event={"ID":"e2d1a4d5-3a08-4d66-8207-ef7e243b2121","Type":"ContainerStarted","Data":"11aeff3e8a2f57c221453940749f736e408a628e93b967727ba55c0ea9ed24e4"} Dec 01 22:55:08 crc kubenswrapper[4857]: I1201 22:55:08.192973 4857 generic.go:334] "Generic (PLEG): container finished" podID="3d27e720-9a0c-419b-958c-a95d40c04774" containerID="1f90f8d700308888a374bf192dec096935c661406767d918fb4aaca52aa6207a" exitCode=0 Dec 01 22:55:08 crc kubenswrapper[4857]: I1201 22:55:08.193020 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nl5jx" event={"ID":"3d27e720-9a0c-419b-958c-a95d40c04774","Type":"ContainerDied","Data":"1f90f8d700308888a374bf192dec096935c661406767d918fb4aaca52aa6207a"} Dec 01 22:55:08 crc kubenswrapper[4857]: I1201 22:55:08.193060 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nl5jx" event={"ID":"3d27e720-9a0c-419b-958c-a95d40c04774","Type":"ContainerStarted","Data":"bfae19c038b5dd05b07f34fed52568393e1a21b7179d34d45afb7df4db960ba4"} Dec 01 22:55:08 crc kubenswrapper[4857]: I1201 22:55:08.194393 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 22:55:09 crc kubenswrapper[4857]: I1201 22:55:09.834627 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:55:09 crc kubenswrapper[4857]: E1201 22:55:09.835385 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:55:10 crc kubenswrapper[4857]: I1201 22:55:10.215080 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nl5jx" event={"ID":"3d27e720-9a0c-419b-958c-a95d40c04774","Type":"ContainerStarted","Data":"65b9a2af140082a36d892808f6ae4e3dbc0ce5a78fa1ce6b823036dd7a10cba3"} Dec 01 22:55:10 crc kubenswrapper[4857]: I1201 22:55:10.221374 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hj2km" event={"ID":"e2d1a4d5-3a08-4d66-8207-ef7e243b2121","Type":"ContainerStarted","Data":"0a3c6cbf6c6e44a81c61e33176cd373514fa6c2867f6d75b0f33628d75f0475c"} Dec 01 22:55:11 crc kubenswrapper[4857]: I1201 22:55:11.232271 4857 generic.go:334] "Generic (PLEG): container finished" podID="e2d1a4d5-3a08-4d66-8207-ef7e243b2121" containerID="0a3c6cbf6c6e44a81c61e33176cd373514fa6c2867f6d75b0f33628d75f0475c" exitCode=0 Dec 01 22:55:11 crc kubenswrapper[4857]: I1201 22:55:11.232351 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hj2km" event={"ID":"e2d1a4d5-3a08-4d66-8207-ef7e243b2121","Type":"ContainerDied","Data":"0a3c6cbf6c6e44a81c61e33176cd373514fa6c2867f6d75b0f33628d75f0475c"} Dec 01 22:55:11 crc kubenswrapper[4857]: I1201 22:55:11.234284 4857 generic.go:334] "Generic (PLEG): container finished" podID="3d27e720-9a0c-419b-958c-a95d40c04774" containerID="65b9a2af140082a36d892808f6ae4e3dbc0ce5a78fa1ce6b823036dd7a10cba3" exitCode=0 Dec 01 22:55:11 crc kubenswrapper[4857]: I1201 22:55:11.234311 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nl5jx" event={"ID":"3d27e720-9a0c-419b-958c-a95d40c04774","Type":"ContainerDied","Data":"65b9a2af140082a36d892808f6ae4e3dbc0ce5a78fa1ce6b823036dd7a10cba3"} Dec 01 22:55:12 crc kubenswrapper[4857]: I1201 22:55:12.249412 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nl5jx" event={"ID":"3d27e720-9a0c-419b-958c-a95d40c04774","Type":"ContainerStarted","Data":"93e182bbe46d0f41413c7131c1333d1128b1d53f8634ccebc7c41bdf609d4377"} Dec 01 22:55:12 crc kubenswrapper[4857]: I1201 22:55:12.255619 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hj2km" event={"ID":"e2d1a4d5-3a08-4d66-8207-ef7e243b2121","Type":"ContainerStarted","Data":"20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2"} Dec 01 22:55:12 crc kubenswrapper[4857]: I1201 22:55:12.282492 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nl5jx" podStartSLOduration=2.761508874 podStartE2EDuration="6.28242013s" podCreationTimestamp="2025-12-01 22:55:06 +0000 UTC" firstStartedPulling="2025-12-01 22:55:08.194660115 +0000 UTC m=+4826.684722432" lastFinishedPulling="2025-12-01 22:55:11.715571381 +0000 UTC m=+4830.205633688" observedRunningTime="2025-12-01 22:55:12.275211243 +0000 UTC m=+4830.765273600" watchObservedRunningTime="2025-12-01 22:55:12.28242013 +0000 UTC m=+4830.772482457" Dec 01 22:55:12 crc kubenswrapper[4857]: I1201 22:55:12.306600 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hj2km" podStartSLOduration=2.780483647 podStartE2EDuration="6.306580111s" podCreationTimestamp="2025-12-01 22:55:06 +0000 UTC" firstStartedPulling="2025-12-01 22:55:08.194106011 +0000 UTC m=+4826.684168328" lastFinishedPulling="2025-12-01 22:55:11.720202465 +0000 UTC m=+4830.210264792" observedRunningTime="2025-12-01 22:55:12.296118445 +0000 UTC m=+4830.786180762" watchObservedRunningTime="2025-12-01 22:55:12.306580111 +0000 UTC m=+4830.796642428" Dec 01 22:55:17 crc kubenswrapper[4857]: I1201 22:55:17.023582 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:17 crc kubenswrapper[4857]: I1201 22:55:17.024394 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:17 crc kubenswrapper[4857]: I1201 22:55:17.119251 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:17 crc kubenswrapper[4857]: I1201 22:55:17.193750 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:17 crc kubenswrapper[4857]: I1201 22:55:17.194233 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:17 crc kubenswrapper[4857]: I1201 22:55:17.284186 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:17 crc kubenswrapper[4857]: I1201 22:55:17.412591 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:17 crc kubenswrapper[4857]: I1201 22:55:17.426097 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:19 crc kubenswrapper[4857]: I1201 22:55:19.184109 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hj2km"] Dec 01 22:55:19 crc kubenswrapper[4857]: I1201 22:55:19.351864 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hj2km" podUID="e2d1a4d5-3a08-4d66-8207-ef7e243b2121" containerName="registry-server" containerID="cri-o://20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2" gracePeriod=2 Dec 01 22:55:19 crc kubenswrapper[4857]: I1201 22:55:19.862567 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nl5jx"] Dec 01 22:55:19 crc kubenswrapper[4857]: I1201 22:55:19.863255 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nl5jx" podUID="3d27e720-9a0c-419b-958c-a95d40c04774" containerName="registry-server" containerID="cri-o://93e182bbe46d0f41413c7131c1333d1128b1d53f8634ccebc7c41bdf609d4377" gracePeriod=2 Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.049156 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.200220 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrmqd\" (UniqueName: \"kubernetes.io/projected/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-kube-api-access-nrmqd\") pod \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\" (UID: \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\") " Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.200340 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-catalog-content\") pod \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\" (UID: \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\") " Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.200391 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-utilities\") pod \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\" (UID: \"e2d1a4d5-3a08-4d66-8207-ef7e243b2121\") " Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.202320 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-utilities" (OuterVolumeSpecName: "utilities") pod "e2d1a4d5-3a08-4d66-8207-ef7e243b2121" (UID: "e2d1a4d5-3a08-4d66-8207-ef7e243b2121"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.205917 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-kube-api-access-nrmqd" (OuterVolumeSpecName: "kube-api-access-nrmqd") pod "e2d1a4d5-3a08-4d66-8207-ef7e243b2121" (UID: "e2d1a4d5-3a08-4d66-8207-ef7e243b2121"). InnerVolumeSpecName "kube-api-access-nrmqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.303725 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrmqd\" (UniqueName: \"kubernetes.io/projected/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-kube-api-access-nrmqd\") on node \"crc\" DevicePath \"\"" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.303765 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.364673 4857 generic.go:334] "Generic (PLEG): container finished" podID="e2d1a4d5-3a08-4d66-8207-ef7e243b2121" containerID="20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2" exitCode=0 Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.364746 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hj2km" event={"ID":"e2d1a4d5-3a08-4d66-8207-ef7e243b2121","Type":"ContainerDied","Data":"20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2"} Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.364820 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hj2km" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.364842 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hj2km" event={"ID":"e2d1a4d5-3a08-4d66-8207-ef7e243b2121","Type":"ContainerDied","Data":"11aeff3e8a2f57c221453940749f736e408a628e93b967727ba55c0ea9ed24e4"} Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.364882 4857 scope.go:117] "RemoveContainer" containerID="20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.412737 4857 scope.go:117] "RemoveContainer" containerID="0a3c6cbf6c6e44a81c61e33176cd373514fa6c2867f6d75b0f33628d75f0475c" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.425983 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2d1a4d5-3a08-4d66-8207-ef7e243b2121" (UID: "e2d1a4d5-3a08-4d66-8207-ef7e243b2121"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.446396 4857 scope.go:117] "RemoveContainer" containerID="385c55244b709a126d5e0176d5747f9754a818b344030371398926a18118f3dc" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.473464 4857 scope.go:117] "RemoveContainer" containerID="20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2" Dec 01 22:55:20 crc kubenswrapper[4857]: E1201 22:55:20.474295 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2\": container with ID starting with 20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2 not found: ID does not exist" containerID="20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.474354 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2"} err="failed to get container status \"20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2\": rpc error: code = NotFound desc = could not find container \"20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2\": container with ID starting with 20c97db26a53e2bfff3c117c09f3e60ac16fd7dc8e84b6a48a9c73ad96ef03a2 not found: ID does not exist" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.474386 4857 scope.go:117] "RemoveContainer" containerID="0a3c6cbf6c6e44a81c61e33176cd373514fa6c2867f6d75b0f33628d75f0475c" Dec 01 22:55:20 crc kubenswrapper[4857]: E1201 22:55:20.474835 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a3c6cbf6c6e44a81c61e33176cd373514fa6c2867f6d75b0f33628d75f0475c\": container with ID starting with 0a3c6cbf6c6e44a81c61e33176cd373514fa6c2867f6d75b0f33628d75f0475c not found: ID does not exist" containerID="0a3c6cbf6c6e44a81c61e33176cd373514fa6c2867f6d75b0f33628d75f0475c" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.474877 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a3c6cbf6c6e44a81c61e33176cd373514fa6c2867f6d75b0f33628d75f0475c"} err="failed to get container status \"0a3c6cbf6c6e44a81c61e33176cd373514fa6c2867f6d75b0f33628d75f0475c\": rpc error: code = NotFound desc = could not find container \"0a3c6cbf6c6e44a81c61e33176cd373514fa6c2867f6d75b0f33628d75f0475c\": container with ID starting with 0a3c6cbf6c6e44a81c61e33176cd373514fa6c2867f6d75b0f33628d75f0475c not found: ID does not exist" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.474897 4857 scope.go:117] "RemoveContainer" containerID="385c55244b709a126d5e0176d5747f9754a818b344030371398926a18118f3dc" Dec 01 22:55:20 crc kubenswrapper[4857]: E1201 22:55:20.475270 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"385c55244b709a126d5e0176d5747f9754a818b344030371398926a18118f3dc\": container with ID starting with 385c55244b709a126d5e0176d5747f9754a818b344030371398926a18118f3dc not found: ID does not exist" containerID="385c55244b709a126d5e0176d5747f9754a818b344030371398926a18118f3dc" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.475304 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"385c55244b709a126d5e0176d5747f9754a818b344030371398926a18118f3dc"} err="failed to get container status \"385c55244b709a126d5e0176d5747f9754a818b344030371398926a18118f3dc\": rpc error: code = NotFound desc = could not find container \"385c55244b709a126d5e0176d5747f9754a818b344030371398926a18118f3dc\": container with ID starting with 385c55244b709a126d5e0176d5747f9754a818b344030371398926a18118f3dc not found: ID does not exist" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.509089 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2d1a4d5-3a08-4d66-8207-ef7e243b2121-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.706372 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hj2km"] Dec 01 22:55:20 crc kubenswrapper[4857]: I1201 22:55:20.722182 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hj2km"] Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.386406 4857 generic.go:334] "Generic (PLEG): container finished" podID="3d27e720-9a0c-419b-958c-a95d40c04774" containerID="93e182bbe46d0f41413c7131c1333d1128b1d53f8634ccebc7c41bdf609d4377" exitCode=0 Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.386895 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nl5jx" event={"ID":"3d27e720-9a0c-419b-958c-a95d40c04774","Type":"ContainerDied","Data":"93e182bbe46d0f41413c7131c1333d1128b1d53f8634ccebc7c41bdf609d4377"} Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.549315 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.642617 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmnxb\" (UniqueName: \"kubernetes.io/projected/3d27e720-9a0c-419b-958c-a95d40c04774-kube-api-access-dmnxb\") pod \"3d27e720-9a0c-419b-958c-a95d40c04774\" (UID: \"3d27e720-9a0c-419b-958c-a95d40c04774\") " Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.642859 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d27e720-9a0c-419b-958c-a95d40c04774-catalog-content\") pod \"3d27e720-9a0c-419b-958c-a95d40c04774\" (UID: \"3d27e720-9a0c-419b-958c-a95d40c04774\") " Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.642996 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d27e720-9a0c-419b-958c-a95d40c04774-utilities\") pod \"3d27e720-9a0c-419b-958c-a95d40c04774\" (UID: \"3d27e720-9a0c-419b-958c-a95d40c04774\") " Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.645189 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d27e720-9a0c-419b-958c-a95d40c04774-utilities" (OuterVolumeSpecName: "utilities") pod "3d27e720-9a0c-419b-958c-a95d40c04774" (UID: "3d27e720-9a0c-419b-958c-a95d40c04774"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.658566 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d27e720-9a0c-419b-958c-a95d40c04774-kube-api-access-dmnxb" (OuterVolumeSpecName: "kube-api-access-dmnxb") pod "3d27e720-9a0c-419b-958c-a95d40c04774" (UID: "3d27e720-9a0c-419b-958c-a95d40c04774"). InnerVolumeSpecName "kube-api-access-dmnxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.727279 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d27e720-9a0c-419b-958c-a95d40c04774-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d27e720-9a0c-419b-958c-a95d40c04774" (UID: "3d27e720-9a0c-419b-958c-a95d40c04774"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.746335 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d27e720-9a0c-419b-958c-a95d40c04774-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.746367 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmnxb\" (UniqueName: \"kubernetes.io/projected/3d27e720-9a0c-419b-958c-a95d40c04774-kube-api-access-dmnxb\") on node \"crc\" DevicePath \"\"" Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.746410 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d27e720-9a0c-419b-958c-a95d40c04774-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 22:55:21 crc kubenswrapper[4857]: I1201 22:55:21.871947 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2d1a4d5-3a08-4d66-8207-ef7e243b2121" path="/var/lib/kubelet/pods/e2d1a4d5-3a08-4d66-8207-ef7e243b2121/volumes" Dec 01 22:55:22 crc kubenswrapper[4857]: I1201 22:55:22.403593 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nl5jx" event={"ID":"3d27e720-9a0c-419b-958c-a95d40c04774","Type":"ContainerDied","Data":"bfae19c038b5dd05b07f34fed52568393e1a21b7179d34d45afb7df4db960ba4"} Dec 01 22:55:22 crc kubenswrapper[4857]: I1201 22:55:22.403683 4857 scope.go:117] "RemoveContainer" containerID="93e182bbe46d0f41413c7131c1333d1128b1d53f8634ccebc7c41bdf609d4377" Dec 01 22:55:22 crc kubenswrapper[4857]: I1201 22:55:22.403738 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nl5jx" Dec 01 22:55:22 crc kubenswrapper[4857]: I1201 22:55:22.443240 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nl5jx"] Dec 01 22:55:22 crc kubenswrapper[4857]: I1201 22:55:22.448292 4857 scope.go:117] "RemoveContainer" containerID="65b9a2af140082a36d892808f6ae4e3dbc0ce5a78fa1ce6b823036dd7a10cba3" Dec 01 22:55:22 crc kubenswrapper[4857]: I1201 22:55:22.457013 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nl5jx"] Dec 01 22:55:22 crc kubenswrapper[4857]: I1201 22:55:22.489272 4857 scope.go:117] "RemoveContainer" containerID="1f90f8d700308888a374bf192dec096935c661406767d918fb4aaca52aa6207a" Dec 01 22:55:23 crc kubenswrapper[4857]: I1201 22:55:23.835152 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:55:23 crc kubenswrapper[4857]: E1201 22:55:23.835826 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:55:23 crc kubenswrapper[4857]: I1201 22:55:23.859283 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d27e720-9a0c-419b-958c-a95d40c04774" path="/var/lib/kubelet/pods/3d27e720-9a0c-419b-958c-a95d40c04774/volumes" Dec 01 22:55:36 crc kubenswrapper[4857]: I1201 22:55:36.836380 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:55:36 crc kubenswrapper[4857]: E1201 22:55:36.837096 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:55:49 crc kubenswrapper[4857]: I1201 22:55:49.835872 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:55:49 crc kubenswrapper[4857]: E1201 22:55:49.837071 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-89689_openshift-machine-config-operator(ac5dfac0-8933-4db8-9cb6-5dca0537d066)\"" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" Dec 01 22:56:01 crc kubenswrapper[4857]: I1201 22:56:01.840405 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 22:56:02 crc kubenswrapper[4857]: I1201 22:56:02.911364 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"e5afdc90ea276d8427ec96233149f13e076b2f35f80416d953de9b16e2d9b328"} Dec 01 22:56:15 crc kubenswrapper[4857]: I1201 22:56:15.067385 4857 generic.go:334] "Generic (PLEG): container finished" podID="51a7c54e-a7cb-49c7-83bc-621241a2dd1a" containerID="a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60" exitCode=0 Dec 01 22:56:15 crc kubenswrapper[4857]: I1201 22:56:15.067500 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qwhq9/must-gather-h5lcq" event={"ID":"51a7c54e-a7cb-49c7-83bc-621241a2dd1a","Type":"ContainerDied","Data":"a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60"} Dec 01 22:56:15 crc kubenswrapper[4857]: I1201 22:56:15.068992 4857 scope.go:117] "RemoveContainer" containerID="a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60" Dec 01 22:56:15 crc kubenswrapper[4857]: I1201 22:56:15.359605 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qwhq9_must-gather-h5lcq_51a7c54e-a7cb-49c7-83bc-621241a2dd1a/gather/0.log" Dec 01 22:56:26 crc kubenswrapper[4857]: I1201 22:56:26.284177 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qwhq9/must-gather-h5lcq"] Dec 01 22:56:26 crc kubenswrapper[4857]: I1201 22:56:26.285839 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-qwhq9/must-gather-h5lcq" podUID="51a7c54e-a7cb-49c7-83bc-621241a2dd1a" containerName="copy" containerID="cri-o://c9424a52371cf8f1c12d017bed32fe71f093641cb58f2bbdaaca0072bca7c936" gracePeriod=2 Dec 01 22:56:26 crc kubenswrapper[4857]: I1201 22:56:26.293290 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qwhq9/must-gather-h5lcq"] Dec 01 22:56:26 crc kubenswrapper[4857]: I1201 22:56:26.776145 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qwhq9_must-gather-h5lcq_51a7c54e-a7cb-49c7-83bc-621241a2dd1a/copy/0.log" Dec 01 22:56:26 crc kubenswrapper[4857]: I1201 22:56:26.776845 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qwhq9/must-gather-h5lcq" Dec 01 22:56:26 crc kubenswrapper[4857]: I1201 22:56:26.903602 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbzv7\" (UniqueName: \"kubernetes.io/projected/51a7c54e-a7cb-49c7-83bc-621241a2dd1a-kube-api-access-jbzv7\") pod \"51a7c54e-a7cb-49c7-83bc-621241a2dd1a\" (UID: \"51a7c54e-a7cb-49c7-83bc-621241a2dd1a\") " Dec 01 22:56:26 crc kubenswrapper[4857]: I1201 22:56:26.904167 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/51a7c54e-a7cb-49c7-83bc-621241a2dd1a-must-gather-output\") pod \"51a7c54e-a7cb-49c7-83bc-621241a2dd1a\" (UID: \"51a7c54e-a7cb-49c7-83bc-621241a2dd1a\") " Dec 01 22:56:26 crc kubenswrapper[4857]: I1201 22:56:26.918279 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a7c54e-a7cb-49c7-83bc-621241a2dd1a-kube-api-access-jbzv7" (OuterVolumeSpecName: "kube-api-access-jbzv7") pod "51a7c54e-a7cb-49c7-83bc-621241a2dd1a" (UID: "51a7c54e-a7cb-49c7-83bc-621241a2dd1a"). InnerVolumeSpecName "kube-api-access-jbzv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.007698 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbzv7\" (UniqueName: \"kubernetes.io/projected/51a7c54e-a7cb-49c7-83bc-621241a2dd1a-kube-api-access-jbzv7\") on node \"crc\" DevicePath \"\"" Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.120062 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51a7c54e-a7cb-49c7-83bc-621241a2dd1a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "51a7c54e-a7cb-49c7-83bc-621241a2dd1a" (UID: "51a7c54e-a7cb-49c7-83bc-621241a2dd1a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.211575 4857 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/51a7c54e-a7cb-49c7-83bc-621241a2dd1a-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.222620 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qwhq9_must-gather-h5lcq_51a7c54e-a7cb-49c7-83bc-621241a2dd1a/copy/0.log" Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.223289 4857 generic.go:334] "Generic (PLEG): container finished" podID="51a7c54e-a7cb-49c7-83bc-621241a2dd1a" containerID="c9424a52371cf8f1c12d017bed32fe71f093641cb58f2bbdaaca0072bca7c936" exitCode=143 Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.223366 4857 scope.go:117] "RemoveContainer" containerID="c9424a52371cf8f1c12d017bed32fe71f093641cb58f2bbdaaca0072bca7c936" Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.223590 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qwhq9/must-gather-h5lcq" Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.274635 4857 scope.go:117] "RemoveContainer" containerID="a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60" Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.395486 4857 scope.go:117] "RemoveContainer" containerID="c9424a52371cf8f1c12d017bed32fe71f093641cb58f2bbdaaca0072bca7c936" Dec 01 22:56:27 crc kubenswrapper[4857]: E1201 22:56:27.396465 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9424a52371cf8f1c12d017bed32fe71f093641cb58f2bbdaaca0072bca7c936\": container with ID starting with c9424a52371cf8f1c12d017bed32fe71f093641cb58f2bbdaaca0072bca7c936 not found: ID does not exist" containerID="c9424a52371cf8f1c12d017bed32fe71f093641cb58f2bbdaaca0072bca7c936" Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.396514 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9424a52371cf8f1c12d017bed32fe71f093641cb58f2bbdaaca0072bca7c936"} err="failed to get container status \"c9424a52371cf8f1c12d017bed32fe71f093641cb58f2bbdaaca0072bca7c936\": rpc error: code = NotFound desc = could not find container \"c9424a52371cf8f1c12d017bed32fe71f093641cb58f2bbdaaca0072bca7c936\": container with ID starting with c9424a52371cf8f1c12d017bed32fe71f093641cb58f2bbdaaca0072bca7c936 not found: ID does not exist" Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.396556 4857 scope.go:117] "RemoveContainer" containerID="a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60" Dec 01 22:56:27 crc kubenswrapper[4857]: E1201 22:56:27.396955 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60\": container with ID starting with a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60 not found: ID does not exist" containerID="a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60" Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.397002 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60"} err="failed to get container status \"a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60\": rpc error: code = NotFound desc = could not find container \"a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60\": container with ID starting with a1576e87f10b4115cc2a37f810be7c98616dba704b3bdabbb23773771ec67c60 not found: ID does not exist" Dec 01 22:56:27 crc kubenswrapper[4857]: I1201 22:56:27.854655 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51a7c54e-a7cb-49c7-83bc-621241a2dd1a" path="/var/lib/kubelet/pods/51a7c54e-a7cb-49c7-83bc-621241a2dd1a/volumes" Dec 01 22:57:25 crc kubenswrapper[4857]: I1201 22:57:25.882747 4857 scope.go:117] "RemoveContainer" containerID="72ca8cf9dd95baf34315474e244185177696ea6c0b3f132fd96b4bfe643a94f0" Dec 01 22:58:02 crc kubenswrapper[4857]: I1201 22:58:02.757407 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="be0a242c-3c02-4d1e-910b-21e74e7c4e43" containerName="galera" probeResult="failure" output="command timed out" Dec 01 22:58:27 crc kubenswrapper[4857]: I1201 22:58:27.806112 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:58:27 crc kubenswrapper[4857]: I1201 22:58:27.806660 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:58:57 crc kubenswrapper[4857]: I1201 22:58:57.806993 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:58:57 crc kubenswrapper[4857]: I1201 22:58:57.807725 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:59:27 crc kubenswrapper[4857]: I1201 22:59:27.807081 4857 patch_prober.go:28] interesting pod/machine-config-daemon-89689 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 22:59:27 crc kubenswrapper[4857]: I1201 22:59:27.807878 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 22:59:27 crc kubenswrapper[4857]: I1201 22:59:27.807952 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-89689" Dec 01 22:59:27 crc kubenswrapper[4857]: I1201 22:59:27.809257 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e5afdc90ea276d8427ec96233149f13e076b2f35f80416d953de9b16e2d9b328"} pod="openshift-machine-config-operator/machine-config-daemon-89689" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 22:59:27 crc kubenswrapper[4857]: I1201 22:59:27.809408 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-89689" podUID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerName="machine-config-daemon" containerID="cri-o://e5afdc90ea276d8427ec96233149f13e076b2f35f80416d953de9b16e2d9b328" gracePeriod=600 Dec 01 22:59:28 crc kubenswrapper[4857]: I1201 22:59:28.762510 4857 generic.go:334] "Generic (PLEG): container finished" podID="ac5dfac0-8933-4db8-9cb6-5dca0537d066" containerID="e5afdc90ea276d8427ec96233149f13e076b2f35f80416d953de9b16e2d9b328" exitCode=0 Dec 01 22:59:28 crc kubenswrapper[4857]: I1201 22:59:28.762658 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerDied","Data":"e5afdc90ea276d8427ec96233149f13e076b2f35f80416d953de9b16e2d9b328"} Dec 01 22:59:28 crc kubenswrapper[4857]: I1201 22:59:28.762916 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-89689" event={"ID":"ac5dfac0-8933-4db8-9cb6-5dca0537d066","Type":"ContainerStarted","Data":"e0b22630c11a56a6a378f1538e1df013c98157ff5ea4b0b637647ea0482d3120"} Dec 01 22:59:28 crc kubenswrapper[4857]: I1201 22:59:28.762992 4857 scope.go:117] "RemoveContainer" containerID="10db3d940a352e11085126611a53958e625da5051b0e7ef316c2aa37f734e0a4" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.162106 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9"] Dec 01 23:00:00 crc kubenswrapper[4857]: E1201 23:00:00.163998 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d1a4d5-3a08-4d66-8207-ef7e243b2121" containerName="registry-server" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.164073 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d1a4d5-3a08-4d66-8207-ef7e243b2121" containerName="registry-server" Dec 01 23:00:00 crc kubenswrapper[4857]: E1201 23:00:00.164143 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d1a4d5-3a08-4d66-8207-ef7e243b2121" containerName="extract-utilities" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.164162 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d1a4d5-3a08-4d66-8207-ef7e243b2121" containerName="extract-utilities" Dec 01 23:00:00 crc kubenswrapper[4857]: E1201 23:00:00.164218 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a7c54e-a7cb-49c7-83bc-621241a2dd1a" containerName="gather" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.164236 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a7c54e-a7cb-49c7-83bc-621241a2dd1a" containerName="gather" Dec 01 23:00:00 crc kubenswrapper[4857]: E1201 23:00:00.164265 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a7c54e-a7cb-49c7-83bc-621241a2dd1a" containerName="copy" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.164280 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a7c54e-a7cb-49c7-83bc-621241a2dd1a" containerName="copy" Dec 01 23:00:00 crc kubenswrapper[4857]: E1201 23:00:00.164314 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d27e720-9a0c-419b-958c-a95d40c04774" containerName="registry-server" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.164334 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d27e720-9a0c-419b-958c-a95d40c04774" containerName="registry-server" Dec 01 23:00:00 crc kubenswrapper[4857]: E1201 23:00:00.164381 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d27e720-9a0c-419b-958c-a95d40c04774" containerName="extract-utilities" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.164399 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d27e720-9a0c-419b-958c-a95d40c04774" containerName="extract-utilities" Dec 01 23:00:00 crc kubenswrapper[4857]: E1201 23:00:00.164444 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d27e720-9a0c-419b-958c-a95d40c04774" containerName="extract-content" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.164459 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d27e720-9a0c-419b-958c-a95d40c04774" containerName="extract-content" Dec 01 23:00:00 crc kubenswrapper[4857]: E1201 23:00:00.164491 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d1a4d5-3a08-4d66-8207-ef7e243b2121" containerName="extract-content" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.164509 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d1a4d5-3a08-4d66-8207-ef7e243b2121" containerName="extract-content" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.165009 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d27e720-9a0c-419b-958c-a95d40c04774" containerName="registry-server" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.165199 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d1a4d5-3a08-4d66-8207-ef7e243b2121" containerName="registry-server" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.165266 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a7c54e-a7cb-49c7-83bc-621241a2dd1a" containerName="copy" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.165304 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a7c54e-a7cb-49c7-83bc-621241a2dd1a" containerName="gather" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.166967 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.173786 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9"] Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.207374 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.207587 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.320878 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f29300f-2488-4746-8c60-93367d671d72-config-volume\") pod \"collect-profiles-29410500-2rcr9\" (UID: \"2f29300f-2488-4746-8c60-93367d671d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.321228 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt9wx\" (UniqueName: \"kubernetes.io/projected/2f29300f-2488-4746-8c60-93367d671d72-kube-api-access-nt9wx\") pod \"collect-profiles-29410500-2rcr9\" (UID: \"2f29300f-2488-4746-8c60-93367d671d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.321288 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f29300f-2488-4746-8c60-93367d671d72-secret-volume\") pod \"collect-profiles-29410500-2rcr9\" (UID: \"2f29300f-2488-4746-8c60-93367d671d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.423596 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f29300f-2488-4746-8c60-93367d671d72-secret-volume\") pod \"collect-profiles-29410500-2rcr9\" (UID: \"2f29300f-2488-4746-8c60-93367d671d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.423734 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f29300f-2488-4746-8c60-93367d671d72-config-volume\") pod \"collect-profiles-29410500-2rcr9\" (UID: \"2f29300f-2488-4746-8c60-93367d671d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.423989 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt9wx\" (UniqueName: \"kubernetes.io/projected/2f29300f-2488-4746-8c60-93367d671d72-kube-api-access-nt9wx\") pod \"collect-profiles-29410500-2rcr9\" (UID: \"2f29300f-2488-4746-8c60-93367d671d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.426254 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f29300f-2488-4746-8c60-93367d671d72-config-volume\") pod \"collect-profiles-29410500-2rcr9\" (UID: \"2f29300f-2488-4746-8c60-93367d671d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.432784 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f29300f-2488-4746-8c60-93367d671d72-secret-volume\") pod \"collect-profiles-29410500-2rcr9\" (UID: \"2f29300f-2488-4746-8c60-93367d671d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.441120 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt9wx\" (UniqueName: \"kubernetes.io/projected/2f29300f-2488-4746-8c60-93367d671d72-kube-api-access-nt9wx\") pod \"collect-profiles-29410500-2rcr9\" (UID: \"2f29300f-2488-4746-8c60-93367d671d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.535293 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:00 crc kubenswrapper[4857]: I1201 23:00:00.863123 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9"] Dec 01 23:00:01 crc kubenswrapper[4857]: I1201 23:00:01.228646 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" event={"ID":"2f29300f-2488-4746-8c60-93367d671d72","Type":"ContainerStarted","Data":"a5c8ebfedadf203ccb4a1095f456f1e47fd11649a94eedb2bd109ebc7e60d855"} Dec 01 23:00:01 crc kubenswrapper[4857]: I1201 23:00:01.228692 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" event={"ID":"2f29300f-2488-4746-8c60-93367d671d72","Type":"ContainerStarted","Data":"3577c3774009522a677e629d8cb9007ca64acba6779bc10b47ae77c72f3f95a5"} Dec 01 23:00:02 crc kubenswrapper[4857]: I1201 23:00:02.242894 4857 generic.go:334] "Generic (PLEG): container finished" podID="2f29300f-2488-4746-8c60-93367d671d72" containerID="a5c8ebfedadf203ccb4a1095f456f1e47fd11649a94eedb2bd109ebc7e60d855" exitCode=0 Dec 01 23:00:02 crc kubenswrapper[4857]: I1201 23:00:02.243098 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" event={"ID":"2f29300f-2488-4746-8c60-93367d671d72","Type":"ContainerDied","Data":"a5c8ebfedadf203ccb4a1095f456f1e47fd11649a94eedb2bd109ebc7e60d855"} Dec 01 23:00:03 crc kubenswrapper[4857]: I1201 23:00:03.638780 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:03 crc kubenswrapper[4857]: I1201 23:00:03.798587 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f29300f-2488-4746-8c60-93367d671d72-secret-volume\") pod \"2f29300f-2488-4746-8c60-93367d671d72\" (UID: \"2f29300f-2488-4746-8c60-93367d671d72\") " Dec 01 23:00:03 crc kubenswrapper[4857]: I1201 23:00:03.799109 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nt9wx\" (UniqueName: \"kubernetes.io/projected/2f29300f-2488-4746-8c60-93367d671d72-kube-api-access-nt9wx\") pod \"2f29300f-2488-4746-8c60-93367d671d72\" (UID: \"2f29300f-2488-4746-8c60-93367d671d72\") " Dec 01 23:00:03 crc kubenswrapper[4857]: I1201 23:00:03.799337 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f29300f-2488-4746-8c60-93367d671d72-config-volume\") pod \"2f29300f-2488-4746-8c60-93367d671d72\" (UID: \"2f29300f-2488-4746-8c60-93367d671d72\") " Dec 01 23:00:03 crc kubenswrapper[4857]: I1201 23:00:03.800052 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f29300f-2488-4746-8c60-93367d671d72-config-volume" (OuterVolumeSpecName: "config-volume") pod "2f29300f-2488-4746-8c60-93367d671d72" (UID: "2f29300f-2488-4746-8c60-93367d671d72"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 23:00:03 crc kubenswrapper[4857]: I1201 23:00:03.805633 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f29300f-2488-4746-8c60-93367d671d72-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2f29300f-2488-4746-8c60-93367d671d72" (UID: "2f29300f-2488-4746-8c60-93367d671d72"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 23:00:03 crc kubenswrapper[4857]: I1201 23:00:03.805744 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f29300f-2488-4746-8c60-93367d671d72-kube-api-access-nt9wx" (OuterVolumeSpecName: "kube-api-access-nt9wx") pod "2f29300f-2488-4746-8c60-93367d671d72" (UID: "2f29300f-2488-4746-8c60-93367d671d72"). InnerVolumeSpecName "kube-api-access-nt9wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 23:00:03 crc kubenswrapper[4857]: I1201 23:00:03.902200 4857 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f29300f-2488-4746-8c60-93367d671d72-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 23:00:03 crc kubenswrapper[4857]: I1201 23:00:03.902234 4857 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f29300f-2488-4746-8c60-93367d671d72-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 23:00:03 crc kubenswrapper[4857]: I1201 23:00:03.902246 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nt9wx\" (UniqueName: \"kubernetes.io/projected/2f29300f-2488-4746-8c60-93367d671d72-kube-api-access-nt9wx\") on node \"crc\" DevicePath \"\"" Dec 01 23:00:04 crc kubenswrapper[4857]: I1201 23:00:04.267026 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" event={"ID":"2f29300f-2488-4746-8c60-93367d671d72","Type":"ContainerDied","Data":"3577c3774009522a677e629d8cb9007ca64acba6779bc10b47ae77c72f3f95a5"} Dec 01 23:00:04 crc kubenswrapper[4857]: I1201 23:00:04.267088 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3577c3774009522a677e629d8cb9007ca64acba6779bc10b47ae77c72f3f95a5" Dec 01 23:00:04 crc kubenswrapper[4857]: I1201 23:00:04.267223 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410500-2rcr9" Dec 01 23:00:04 crc kubenswrapper[4857]: I1201 23:00:04.381489 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l"] Dec 01 23:00:04 crc kubenswrapper[4857]: I1201 23:00:04.391195 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410455-sct6l"] Dec 01 23:00:05 crc kubenswrapper[4857]: I1201 23:00:05.855784 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="891266c7-ed7c-48c1-bbe9-2fba0bb9da3a" path="/var/lib/kubelet/pods/891266c7-ed7c-48c1-bbe9-2fba0bb9da3a/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515113416777024461 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015113417000017352 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113404454016506 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113404455015457 5ustar corecore